Compare commits

..

222 Commits

Author SHA1 Message Date
869137ae23 Merge pull request 'feat: add Ezra MemPalace operator bundle (#570)' (#837) from fix/570 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Failing after 29s
Smoke Test / smoke (push) Failing after 27s
2026-04-21 15:40:15 +00:00
2beda70e9c Merge pull request 'feat: add codebase genome status rollup (#665)' (#817) from fix/665 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:39:56 +00:00
13b25a5fbd Merge pull request 'docs: refresh the-playground genome analysis (#671)' (#813) from fix/671 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:39:49 +00:00
f9965180af Merge pull request 'docs: verify grounded slice for #545' (#782) from fix/545 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:39:32 +00:00
e82e9d9e48 Merge pull request 'docs: verify #567 already implemented on main' (#781) from fix/567 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:39:25 +00:00
911e860dce Merge pull request 'fix: refresh burn lane audit for #662' (#780) from fix/662 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:39:20 +00:00
a391845126 Merge pull request 'feat: add truck battery disconnect install packet (#528)' (#779) from fix/528 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:39:19 +00:00
3b923ab2b3 Merge pull request 'feat: audit trail — local logging of inputs, sources, confidence (#794)' (#799) from feat/794-audit-trail into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:36:52 +00:00
6392cc7a78 Merge pull request 'docs: verify #648 already implemented' (#818) from fix/648 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:36:38 +00:00
2d733566f2 Merge pull request 'fix(#715): Fix smoke workflow JSON parse step and add pytest coverage' (#823) from sprint/issue-822 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:36:29 +00:00
6e8de07236 Merge pull request 'fix: docs: refresh turboquant codebase genome for #679' (#828) from sprint/issue-827 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:36:19 +00:00
ffbf12b8dc Merge pull request 'feat: add compounding-intelligence genome analysis (#676)' (#833) from fix/676 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:36:15 +00:00
f2bb275879 Merge pull request 'feat: ground phase-6 network state (#553)' (#838) from fix/553 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:36:05 +00:00
e6c9a58167 Merge pull request 'docs: verify #675 already implemented' (#834) from fix/675 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:34:03 +00:00
c67e0b535f Merge pull request 'docs: refresh the-door genome analysis (#673)' (#835) from fix/673 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:34:01 +00:00
00039d496c Merge pull request 'fix(#551): sovereignty audit tool — measure cloud dependencies' (#839) from fix/551 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:33:51 +00:00
6cc15a1511 Merge pull request 'fix: correct sparse baseline handling for #749' (#787) from fix/749 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:31:54 +00:00
3a201808bc Merge pull request 'feat: crisis detector for SOUL.md compliance — 988 Lifeline integration (#791)' (#801) from fix/791-crisis-detector into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:31:44 +00:00
112d670a1b Merge pull request 'feat: grounding before generation - retrieval is not a feature (#792)' (#804) from fix/792-grounding into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:31:38 +00:00
2dcd3ba988 Merge pull request 'docs: verify #680 already implemented' (#807) from fix/680 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:31:35 +00:00
2fd1b80792 Merge pull request '[OPS] timmy-home backlog reduced from 220 to 50 — triage cadence needed' (#829) from fix/685 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:31:13 +00:00
11ff6138cc Merge pull request 'feat: refresh evennia-local-world genome and tests (#677)' (#832) from fix/677 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:31:07 +00:00
d97174f338 Merge pull request 'fix(#548): Phase 1 Survival — real metrics, gap analysis, snapshot' (#840) from fix/548 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:30:52 +00:00
b432ef0e25 Merge pull request 'fix(#544): tests for Bezalel Gemma 4 GPU provisioning scaffold' (#841) from fix/544 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:30:48 +00:00
4c772fe5be Merge pull request 'feat: add fleet progression status report (#547)' (#842) from fix/547 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:30:46 +00:00
b4d4681a4b Merge pull request 'docs: verify epic slice for #582 on main' (#789) from fix/582 into main 2026-04-21 15:25:56 +00:00
96d4d1fb3b Merge pull request 'docs: verify #693 already implemented on main' (#788) from fix/693 into main
Some checks failed
Self-Healing Smoke / self-healing-smoke (push) Has been cancelled
Smoke Test / smoke (push) Has been cancelled
2026-04-21 15:25:52 +00:00
6790170732 Add tests/test_bezalel_gemma4.py
Some checks failed
Smoke Test / smoke (pull_request) Failing after 33s
2026-04-21 11:33:36 +00:00
Alexander Whitestone
ef6a729c32 feat: add fleet progression status report (#547)
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 50s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 30s
Smoke Test / smoke (pull_request) Failing after 19s
Agent PR Gate / report (pull_request) Successful in 19s
2026-04-21 07:33:04 -04:00
Alexander Whitestone
7926c74cb6 test: cover fleet progression status report (#547) 2026-04-21 07:29:34 -04:00
60d876c12d fix(#548): complete Phase 1 survival doc with real metrics, gap analysis, run instructions
Some checks failed
Smoke Test / smoke (pull_request) Failing after 28s
2026-04-21 11:28:31 +00:00
08e5356b01 fix(#548): add Phase 1 survival snapshot with current fleet state 2026-04-21 11:27:59 +00:00
8b02ae03ca Add scripts/sovereignty_audit.py
Some checks failed
Smoke Test / smoke (pull_request) Failing after 23s
2026-04-21 11:27:38 +00:00
Alexander Whitestone
e752caa9a7 feat: ground phase-6 network state (#553)
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 25s
Agent PR Gate / gate (pull_request) Failing after 46s
Smoke Test / smoke (pull_request) Failing after 20s
Agent PR Gate / report (pull_request) Successful in 24s
2026-04-21 07:23:21 -04:00
Alexander Whitestone
5c8ba43dbf feat: add Ezra MemPalace operator bundle (#570)
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 31s
Agent PR Gate / gate (pull_request) Failing after 1m2s
Smoke Test / smoke (pull_request) Failing after 28s
Agent PR Gate / report (pull_request) Successful in 19s
2026-04-21 07:22:18 -04:00
Alexander Whitestone
81e5fa4a54 test: cover phase-6 network grounding (#553) 2026-04-21 07:21:22 -04:00
Alexander Whitestone
cf461ec99f wip: extend Ezra MemPalace regression coverage (#570) 2026-04-21 07:20:25 -04:00
Alexander Whitestone
4dfa001b9a docs: refresh the-door genome analysis (#673)
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 27s
Smoke Test / smoke (pull_request) Failing after 27s
Agent PR Gate / gate (pull_request) Failing after 39s
Agent PR Gate / report (pull_request) Successful in 9s
2026-04-21 03:55:52 -04:00
Alexander Whitestone
0173ed67e2 wip: strengthen the-door genome regression for #673 2026-04-21 03:51:28 -04:00
Alexander Whitestone
16fcabb5fc docs: verify #675 already implemented
Some checks failed
Smoke Test / smoke (pull_request) Failing after 25s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 32s
Agent PR Gate / gate (pull_request) Failing after 39s
Agent PR Gate / report (pull_request) Successful in 8s
2026-04-21 03:41:35 -04:00
Alexander Whitestone
72159c1714 wip: add the-testament genome verification regression for #675 2026-04-21 03:40:26 -04:00
Alexander Whitestone
0626a3fc33 feat: add compounding-intelligence genome analysis (#676)
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 25s
Smoke Test / smoke (pull_request) Failing after 30s
Agent PR Gate / gate (pull_request) Failing after 38s
Agent PR Gate / report (pull_request) Successful in 10s
2026-04-21 03:27:24 -04:00
Alexander Whitestone
98f861b713 wip: add compounding-intelligence genome regression for #676 2026-04-21 03:23:05 -04:00
Alexander Whitestone
3250c6d124 feat: refresh evennia-local-world genome and tests (#677)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 30s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 37s
Agent PR Gate / gate (pull_request) Failing after 45s
Agent PR Gate / report (pull_request) Successful in 11s
2026-04-21 03:10:50 -04:00
Alexander Whitestone
ab050629fc wip: add evennia-local-world genome regression for #677 2026-04-21 03:06:33 -04:00
Alexander Whitestone
1bce98a761 fix: #685
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 48s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 23s
Smoke Test / smoke (pull_request) Failing after 22s
Agent PR Gate / report (pull_request) Successful in 13s
- Add weekly triage cadence for timmy-home
- Update scripts/backlog_triage.py with weekly analysis
- Add docs/weekly-triage-cadence.md with documentation

Addresses issue #685: [OPS] timmy-home backlog reduced from 220 to 50 — triage cadence needed

Features:
1. Weekly backlog analysis
2. Report generation
3. Cron entry generation
4. JSON output for automation

Usage:
- python scripts/backlog_triage.py --analyze
- python scripts/backlog_triage.py --report
- python scripts/backlog_triage.py --cron
- python scripts/backlog_triage.py --json

Metrics tracked:
- Total open issues
- Unassigned issues
- Unlabeled issues
- Stale issues (>30 days)
- Batch-pipeline issues

Integration:
- Weekly cron job
- Morning report
- CI/CD workflow
2026-04-20 22:46:50 -04:00
Alexander Whitestone
cb269347cc fix: docs: refresh turboquant codebase genome for #679 (closes #827)
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 38s
Agent PR Gate / gate (pull_request) Failing after 50s
Smoke Test / smoke (pull_request) Failing after 24s
Agent PR Gate / report (pull_request) Successful in 17s
2026-04-20 21:42:33 -04:00
Alexander Whitestone
47f136e9ab fix: fix(#715): Fix smoke workflow JSON parse step and add pytest coverage (closes #822)
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 22s
Agent PR Gate / gate (pull_request) Failing after 48s
Smoke Test / smoke (pull_request) Failing after 28s
Agent PR Gate / report (pull_request) Successful in 11s
2026-04-20 20:14:15 -04:00
Alexander Whitestone
b615013e63 docs: verify #648 already implemented
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 16s
Smoke Test / smoke (pull_request) Failing after 17s
Agent PR Gate / gate (pull_request) Failing after 29s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-18 15:26:57 -04:00
Alexander Whitestone
e4e63cdbb7 wip: add issue 648 verification doc test 2026-04-18 15:26:30 -04:00
Alexander Whitestone
0b18f106b9 feat: add codebase genome status rollup (#665)
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 17s
Smoke Test / smoke (pull_request) Failing after 18s
Agent PR Gate / gate (pull_request) Failing after 28s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-18 15:16:05 -04:00
Alexander Whitestone
2f3138db6d wip: add codebase genome status regression tests 2026-04-18 15:16:05 -04:00
Alexander Whitestone
c2aed12464 docs: refresh the-playground genome analysis (#671)
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 24s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 11s
Smoke Test / smoke (pull_request) Failing after 17s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-17 03:09:43 -04:00
Alexander Whitestone
d9221d5cd6 wip: refresh the-playground genome regression test 2026-04-17 03:09:43 -04:00
Alexander Whitestone
23e62ed5d3 docs: verify issue 680 already implemented
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 17s
Agent PR Gate / gate (pull_request) Failing after 40s
Smoke Test / smoke (pull_request) Failing after 18s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-17 02:10:32 -04:00
Alexander Whitestone
29e3a3f06c test: add issue 680 verification doc regression 2026-04-17 02:08:28 -04:00
Alexander Whitestone
55c8100b8f feat: grounding before generation - retrieval is not a feature (#792)
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 21s
Agent PR Gate / gate (pull_request) Failing after 22s
Smoke Test / smoke (pull_request) Failing after 17s
Agent PR Gate / report (pull_request) Has been cancelled
SOUL.md compliance: 'When I have verified sources, I must consult them
before I generate from pattern alone. Retrieval is not a feature.
It is the primary mechanism by which I avoid lying.'

scripts/grounding.py:
  GroundingLayer with ground() - queries memory files + context before generation
  GroundingResult with grounded flag, confidence, sources, hedging indicator
  format_sources() for display
  Searches memory/*.md and provided context text

Tests: 6 passing
2026-04-17 01:52:48 -04:00
Alexander Whitestone
1f92fb0480 feat: source distinction - I think vs I know (#793)
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 18s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 17s
Smoke Test / smoke (pull_request) Failing after 18s
Agent PR Gate / report (pull_request) Has been cancelled
SOUL.md compliance: 'Every claim I make comes from one of two places:
a verified source I can point to, or my own pattern-matching.'

scripts/source_distinction.py:
  SourceType enum: VERIFIED, INFERRED, STATED, UNKNOWN
  Claim dataclass with source_type, source_ref, confidence, hedging
  AnnotatedResponse with render() and format_for_display()
  Helper functions: verified(), inferred(), stated()
  source_distinction_check() - hedging word detection

Tests: 9 passing
2026-04-17 01:44:06 -04:00
ed179d5e75 test: crisis detector test suite (#791)\n\nAdversarial, multilingual, edge cases.\n>95% recall target on explicit ideation.\nZero false positives on normal conversation.
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 22s
Agent PR Gate / gate (pull_request) Failing after 49s
Smoke Test / smoke (pull_request) Failing after 19s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-17 05:35:30 +00:00
5c52bd83f6 feat: crisis detector for SOUL.md compliance (#791)\n\nDetects suicidal/hopeless language in user input.\nRoutes to 988 Lifeline. Shares the gospel.\nNever computes value of human life.\nCloses #791 2026-04-17 05:34:33 +00:00
Alexander Whitestone
a39f4fb1ab feat: audit trail - local logging of inputs, sources, confidence (#794)
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 21s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 18s
Smoke Test / smoke (pull_request) Failing after 18s
Agent PR Gate / report (pull_request) Has been cancelled
SOUL.md compliance: 'Every response I generate should be logged locally
with the inputs that produced it, the sources I consulted, and the
confidence assessment I made.'

scripts/audit_trail.py:
  AuditTrail class with log_response(), query(), get_stats(), get_by_session()
  Content-addressed IDs, JSONL storage, local-only (never sent anywhere)
  User can query: 'why did you say X?' -> search trail for matching entries

tests/test_audit_trail.py: 7 tests passing
2026-04-17 01:34:15 -04:00
ec444d0749 feat: audit trail tests (#794)
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 38s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 20s
Smoke Test / smoke (pull_request) Failing after 20s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-17 05:30:35 +00:00
315c36a35d feat: tests/timmy init (#794) 2026-04-17 05:30:34 +00:00
739281217d feat: tests init (#794) 2026-04-17 05:30:33 +00:00
36c5a44dff feat: audit trail — SOUL.md honesty requirement (#794) 2026-04-17 05:29:02 +00:00
793497e277 feat: init timmy package (#794) 2026-04-17 05:29:00 +00:00
Alexander Whitestone
08ceb99cac docs: verify epic slice for #582 on main
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 13s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 5s
Smoke Test / smoke (pull_request) Failing after 5s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-17 00:36:08 -04:00
Alexander Whitestone
6a8d8d8392 test: define verification evidence for #582 2026-04-17 00:34:51 -04:00
Alexander Whitestone
6c9ef6b4ef docs: verify #693 already implemented on main
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 13s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 6s
Smoke Test / smoke (pull_request) Failing after 6s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-17 00:29:10 -04:00
Alexander Whitestone
6a56e39fa3 test: define verification evidence for #693 2026-04-17 00:27:58 -04:00
Alexander Whitestone
909b88af56 fix: use prior active window baseline for #749
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 12s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 5s
Smoke Test / smoke (pull_request) Failing after 5s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-17 00:19:50 -04:00
Alexander Whitestone
f9f342cee7 test: capture sparse baseline fallback for #749 2026-04-17 00:17:21 -04:00
Alexander Whitestone
1d2fbc747b docs: verify grounded slice for #545
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 11s
Agent PR Gate / gate (pull_request) Failing after 23s
Smoke Test / smoke (pull_request) Failing after 10s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-16 23:57:26 -04:00
Alexander Whitestone
5ad0adee65 test: define verification evidence for #545 2026-04-16 23:56:06 -04:00
Alexander Whitestone
3d57f42adc docs: verify #567 already implemented on main
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 14s
Agent PR Gate / gate (pull_request) Failing after 27s
Smoke Test / smoke (pull_request) Failing after 10s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-16 23:50:54 -04:00
Alexander Whitestone
bb24a9ab4c test: define verification evidence for #567 2026-04-16 23:50:01 -04:00
Alexander Whitestone
6cbb9a98e1 fix: refresh burn lane audit for #662
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 48s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 8s
Smoke Test / smoke (pull_request) Failing after 6s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-16 23:43:24 -04:00
Alexander Whitestone
0716234d00 feat: add truck battery disconnect install packet (#528)
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 32s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 12s
Smoke Test / smoke (pull_request) Failing after 8s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-16 22:17:42 -04:00
Alexander Whitestone
a8121aa4e9 test: define battery disconnect packet acceptance for #528 2026-04-16 22:10:35 -04:00
37a08f45b8 Merge pull request 'docs: MemPalace v3.0.0 integration — before/after evaluation (#568)' (#764) from fix/568-mempalace-evaluation into main
Merge PR #764: docs: MemPalace v3.0.0 integration — before/after evaluation (#568)
2026-04-17 01:46:41 +00:00
9c420127be Merge pull request 'docs: add the-nexus genome analysis (#672)' (#763) from fix/672 into main
Merge PR #763: docs: add the-nexus genome analysis (#672)
2026-04-17 01:46:38 +00:00
13eea2ce44 Merge pull request 'feat: Codebase Genome for burn-fleet — 112-pane dispatch infrastructure (#681)' (#762) from fix/681-burn-fleet-genome into main
Merge PR #762: feat: Codebase Genome for burn-fleet — 112-pane dispatch infrastructure (#681)
2026-04-17 01:46:36 +00:00
Alexander Whitestone
5c2cf06f57 fix: Evennia settings for Bezalel VPS (#534)
Some checks failed
Agent PR Gate / gate (pull_request) Has been cancelled
Agent PR Gate / report (pull_request) Has been cancelled
Self-Healing Smoke / self-healing-smoke (pull_request) Has been cancelled
Smoke Test / smoke (pull_request) Has been cancelled
Fix script removes bad port tuples (None values) that crash Evennia
Twisted port binding, replaces with correct format, re-migrates DB,
creates superuser, and starts Evennia.

Run via SSH:
  ssh root@104.131.15.18 'bash -s' < scripts/fix_evennia_settings.sh

Fixes:
  - WEBSERVER_PORTS: (4101, None) -> (4001, 0.0.0.0)
  - TELNET_PORTS: None -> (4000, 0.0.0.0)
  - SERVERNAME set to bezalel_world
  - DB cleaned and re-migrated
  - Superuser Timmy created
2026-04-16 01:59:25 -04:00
Alexander Whitestone
4fd78ace44 feat: Bezalel Evennia world builder (#536)
Some checks failed
Agent PR Gate / gate (pull_request) Has been cancelled
Agent PR Gate / report (pull_request) Has been cancelled
Self-Healing Smoke / self-healing-smoke (pull_request) Has been cancelled
Smoke Test / smoke (pull_request) Has been cancelled
2026-04-16 01:52:44 -04:00
Alexander Whitestone
b8b8bb65fd feat: GENOME.md template + single-repo analyzer (#666)
Some checks failed
Agent PR Gate / gate (pull_request) Has been cancelled
Agent PR Gate / report (pull_request) Has been cancelled
Self-Healing Smoke / self-healing-smoke (pull_request) Has been cancelled
Smoke Test / smoke (pull_request) Has been cancelled
Template and analyzer for the Codebase Genome batch (issues #667-683).

templates/GENOME-template.md:
  Reusable template with placeholders for repo name, overview,
  architecture, entry points, data flow, abstractions, API surface,
  test coverage, security, and design decisions.

scripts/genome_analyzer.py:
  Auto-generates GENOME.md skeleton from a codebase scan:
  - File counts by extension
  - Directory structure (depth 2)
  - Entry point detection (main.py, scripts/, Makefile, etc.)
  - Test file enumeration
  - README first-paragraph extraction

Usage:
  python3 scripts/genome_analyzer.py /path/to/repo --output GENOME.md
  python3 scripts/genome_analyzer.py /path/to/repo --dry-run
2026-04-16 01:19:41 -04:00
8e86b8c3de docs: MemPalace v3.0.0 integration evaluation (#568)
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 13s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 13s
Smoke Test / smoke (pull_request) Failing after 11s
Agent PR Gate / report (pull_request) Has been cancelled
Before/after evaluation report for MemPalace integration.
Key findings:
- 96.6% R@5 with zero API calls
- +34% retrieval boost from palace structure
- 210-token wake-up context
- MCP compatible, fully local

Recommendation: Integrate as primary memory layer.

Closes #568.
2026-04-16 00:35:22 -04:00
ff7ea2d45e feat: Codebase Genome for burn-fleet (#681)
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 43s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 34s
Smoke Test / smoke (pull_request) Failing after 13m50s
Agent PR Gate / report (pull_request) Has been cancelled
Complete GENOME.md for burn-fleet (autonomous dispatch infra):
- Project overview: 112 panes, 96 workers across Mac + VPS
- Architecture diagram (ASCII)
- Lane routing table (8 repos → windows)
- Agent name registry (48 mythological names)
- Entry points and design decisions
- Scaling instructions
- Sovereignty assessment

Repo 14/16. Closes #681.
2026-04-16 00:29:30 -04:00
Alexander Whitestone
5c7ba5475f docs: add the-nexus genome analysis (refs #672)
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 18s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 8s
Smoke Test / smoke (pull_request) Failing after 7s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-16 00:29:22 -04:00
Alexander Whitestone
5d7b26858e test: define the-nexus genome acceptance for #672 2026-04-16 00:20:16 -04:00
Alexander Whitestone
c4a04ecb6d Merge PR #710 2026-04-16 00:02:58 -04:00
Alexander Whitestone
ed8c60ecd8 Merge PR #711 2026-04-16 00:02:55 -04:00
Alexander Whitestone
2a8beaf5e3 Merge PR #712 2026-04-16 00:02:49 -04:00
Alexander Whitestone
e2ff2a335f Merge PR #713 2026-04-16 00:02:47 -04:00
Alexander Whitestone
9c1f24e1e9 Merge PR #714 2026-04-16 00:02:45 -04:00
Alexander Whitestone
300ded011a Merge PR #717 2026-04-16 00:02:42 -04:00
Alexander Whitestone
a8a17c1bf4 Merge PR #718 2026-04-16 00:02:40 -04:00
Alexander Whitestone
d98be5bb64 Merge PR #719 2026-04-16 00:02:37 -04:00
Alexander Whitestone
d340c58409 Merge PR #720 2026-04-16 00:02:34 -04:00
Alexander Whitestone
df8b0b32b0 Merge PR #721 2026-04-16 00:02:30 -04:00
Alexander Whitestone
e450713e8e Merge PR #722 2026-04-16 00:02:27 -04:00
Alexander Whitestone
e5a7cff6fe Merge PR #723 2026-04-16 00:02:21 -04:00
Alexander Whitestone
9c3ca942f5 Merge PR #724 2026-04-16 00:02:19 -04:00
Alexander Whitestone
1607b458a4 Merge PR #725 2026-04-16 00:02:16 -04:00
Alexander Whitestone
87477d2447 Merge PR #726 2026-04-16 00:02:14 -04:00
Alexander Whitestone
6968125123 Merge PR #727 2026-04-16 00:02:10 -04:00
Alexander Whitestone
b9c7f7049c Merge PR #728 2026-04-16 00:02:06 -04:00
Alexander Whitestone
1cd63847d4 Merge PR #729 2026-04-16 00:01:58 -04:00
Alexander Whitestone
0ff4f4b023 Merge PR #730 2026-04-16 00:01:52 -04:00
Alexander Whitestone
6a35135cfb Merge PR #731 2026-04-16 00:01:48 -04:00
Alexander Whitestone
0ac749fad4 Merge PR #732 2026-04-16 00:01:41 -04:00
Alexander Whitestone
455dbab287 Merge PR #733 2026-04-16 00:01:35 -04:00
Alexander Whitestone
347d996a32 Merge PR #734 2026-04-16 00:01:29 -04:00
Alexander Whitestone
843f4f422f Merge PR #736 2026-04-16 00:01:22 -04:00
Alexander Whitestone
efc727c5c8 Merge PR #737 2026-04-16 00:01:15 -04:00
Alexander Whitestone
88a47ce77f Merge PR #742 2026-04-16 00:01:09 -04:00
Alexander Whitestone
c07ed5f218 Merge PR #743 2026-04-16 00:01:03 -04:00
Alexander Whitestone
1ef52f8922 Merge PR #744 2026-04-16 00:00:57 -04:00
Alexander Whitestone
5851b0e1ad Merge PR #746 2026-04-16 00:00:50 -04:00
Alexander Whitestone
df0980509d Merge PR #747 2026-04-16 00:00:45 -04:00
Alexander Whitestone
29b1336bf3 Merge PR #753 2026-04-16 00:00:27 -04:00
Alexander Whitestone
c44b0b460e Merge PR #755 2026-04-16 00:00:22 -04:00
Alexander Whitestone
6e79ce633e Merge PR #756 2026-04-16 00:00:19 -04:00
Alexander Whitestone
588b038132 Merge PR #757 2026-04-16 00:00:16 -04:00
Alexander Whitestone
08777276bd Merge PR #758 2026-04-16 00:00:14 -04:00
Alexander Whitestone
359e4b4e7c Merge PR #759 2026-04-16 00:00:09 -04:00
Alexander Whitestone
4b7c133c3e Merge PR #686 2026-04-15 23:59:45 -04:00
Alexander Whitestone
428bb32a30 Merge PR #687 2026-04-15 23:59:43 -04:00
Alexander Whitestone
d1f7a2e63d Merge PR #689 2026-04-15 23:59:38 -04:00
Alexander Whitestone
a7762aabf2 Merge PR #690 2026-04-15 23:59:35 -04:00
Alexander Whitestone
b6519aa939 Merge PR #695 2026-04-15 23:59:32 -04:00
Alexander Whitestone
6526e53579 Merge PR #696 2026-04-15 23:59:30 -04:00
Alexander Whitestone
9553ff5c14 Merge PR #697 2026-04-15 23:59:26 -04:00
Alexander Whitestone
24036f3ed9 Merge PR #698 2026-04-15 23:59:23 -04:00
Alexander Whitestone
6527462727 Merge PR #699 2026-04-15 23:59:20 -04:00
Alexander Whitestone
1b9e8184c5 Merge PR #700 2026-04-15 23:59:17 -04:00
Alexander Whitestone
2b48cd0e42 Merge PR #701 2026-04-15 23:59:15 -04:00
Alexander Whitestone
612b8ac068 Merge PR #702 2026-04-15 23:59:11 -04:00
Alexander Whitestone
6798d68f69 Merge PR #703 2026-04-15 23:59:07 -04:00
Alexander Whitestone
54c69b7d8b Merge PR #704 2026-04-15 23:59:03 -04:00
Alexander Whitestone
12bfe5d1bc Merge PR #705 2026-04-15 23:58:58 -04:00
Alexander Whitestone
2f93f829ee Merge PR #706 2026-04-15 23:58:56 -04:00
Alexander Whitestone
7f28ddc4da Merge PR #707 2026-04-15 23:58:54 -04:00
Alexander Whitestone
d50e236b73 Merge PR #708 2026-04-15 23:58:52 -04:00
Alexander Whitestone
c13dbbbcda Merge PR #709 2026-04-15 23:58:47 -04:00
c46bec5d6b Merge pull request 'feat: sovereign DNS management via Cloudflare API (#692)' (#761) from fix/692-sovereign-dns into main 2026-04-16 03:50:58 +00:00
5e1aeb7b5b Merge pull request 'feat: cross-repo QA automation script (#691)' (#760) from fix/691 into main 2026-04-16 03:49:07 +00:00
Alexander Whitestone
ce3da2dbc4 feat: sovereign DNS management via Cloudflare API (#692)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 2m25s
DNS record management script with Cloudflare API integration:

Operations:
  python3 scripts/dns-manager.py list --zone example.com
  python3 scripts/dns-manager.py add --zone example.com --name forge --ip 1.2.3.4
  python3 scripts/dns-manager.py update --zone example.com --name forge --ip 5.6.7.8
  python3 scripts/dns-manager.py delete --zone example.com --name forge
  python3 scripts/dns-manager.py sync --zone example.com --config dns-records.yaml

Features:
- Cloudflare API v4 integration (stdlib only, no deps)
- Zone auto-resolution from domain name
- Sync from YAML config (add missing, update changed)
- API token from CLOUDFLARE_API_TOKEN env or ~/.config/cloudflare/token
- Fleet DNS records: forge, bezalel, allegro subdomains

Also: dns-records.yaml with current fleet domain mappings.
2026-04-15 23:47:32 -04:00
5b0438f2f5 feat: cross-repo QA automation script (#691)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 28s
2026-04-16 02:12:17 +00:00
2c31ae8972 docs: predictive resource allocation — operator guide (#749)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 19s
Closes #749
2026-04-16 01:51:19 +00:00
Alexander Whitestone
e6279b856a Merge remote-tracking branch 'origin/main' into fix/680 2026-04-15 21:50:46 -04:00
a76e83439c test: predictive resource allocator (#749)
25 tests covering:
- Timestamp parsing, rate computation, caller analysis
- Heartbeat risk detection, demand prediction
- Posture determination, full forecast pipeline
- Markdown output

Closes #749
2026-04-16 01:49:51 +00:00
a14a233626 feat: predictive resource allocation — forecast fleet demand (#749)
Analyzes historical metrics and heartbeat logs to predict
workload surges and recommend pre-provisioning actions.

Closes #749
2026-04-16 01:47:48 +00:00
fa450d8b19 docs: LAB-004 solar deployment guide (#529)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 20s
Complete deployment guide for 600W cabin solar array:
- ASCII wiring diagram (panels → controller → batteries → inverter → AC)
- Parts list with estimated costs ($285-405, under $500 budget)
- 2S2P panel wiring configuration (36V to controller)
- 4-battery parallel bank layout
- Step-by-step installation checklist
- Safety notes and estimated runtime

Physical hardware task — documentation only.

Closes #529.
2026-04-15 21:45:16 -04:00
a251d3b75d feat: test suite generator — fill coverage gaps (#667)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 23s
2026-04-16 01:39:58 +00:00
601c5fe267 Merge pull request 'research: Long Context vs RAG Decision Framework (backlog item #4.3)' (#750) from research/long-context-vs-rag into main 2026-04-16 01:39:55 +00:00
Alexander Whitestone
b3dd906805 docs: add fleet-ops genome analysis (#680) 2026-04-15 21:36:08 -04:00
c9122809c8 feat: Codebase Genome for evennia-local-world (#677)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 14s
Complete GENOME.md for Evennia local world (MUD layer):
- Project overview: MUD + MemPalace + Nexus visualization
- Architecture diagram (Mermaid)
- Key components, event protocol, commands
- Test coverage, sovereignty assessment

Source: the-nexus/nexus/evennia_mempalace/
Repo 10/16. Closes #677.
2026-04-15 21:26:05 -04:00
Alexander Whitestone
58749454e0 ops: audit burn lane empty report for #662
Some checks failed
Smoke Test / smoke (pull_request) Failing after 33s
2026-04-15 21:23:59 -04:00
Alexander Whitestone
3ada0c10c8 test: define burn lane empty audit acceptance for #662 2026-04-15 21:20:44 -04:00
372ffa3fdf feat: Codebase Genome for the-nexus (#672)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 17s
Complete GENOME.md for the-nexus (3D world + MUD + memory):
- Project overview: Three.js + Evennia + MemPalace
- Architecture diagram (Mermaid)
- 10 key subsystems documented
- Entry points (browser, server, electron, deploy)
- MemPalace system breakdown
- Evennia integration details
- Configuration and documentation index
- Sovereignty assessment

Repo 5/16. Closes #672.
2026-04-15 21:04:59 -04:00
Alexander Whitestone
8a0ffc190d docs: add the-beacon genome analysis (refs #674) 2026-04-15 21:03:00 -04:00
f684b0deb8 feat: Codebase Genome for turboquant (#679)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 17s
Complete GENOME.md for turboquant (KV cache compression):
- Project overview: PolarQuant + QJL = 3.5bit/channel
- Architecture diagram (Mermaid)
- Entry points and data flow
- Key abstractions (encode/decode/Metal shaders)
- File index (~660 LOC)
- Upstream source repos
- Test coverage
- Sovereignty assessment

Repo 12/16. Closes #679.
2026-04-15 20:59:55 -04:00
Alexander Whitestone
f337cff98e test: define the-beacon genome acceptance for #674 2026-04-15 20:56:39 -04:00
f76c8187cf docs: triage cadence report for #685
Some checks failed
Smoke Test / smoke (pull_request) Failing after 15s
Backlog reduced from 220 to 50. Report documents triage
cadence needed to maintain healthy backlog.

- Daily: 5 min new issue check
- Weekly: 15 min full sweep
- Monthly: 30 min audit

Closes #685.
2026-04-15 20:55:22 -04:00
6222b18a38 research: Long Context vs RAG Decision Framework (backlog item #4.3)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 18s
Highest-ratio research item (Impact:4, Effort:1, Ratio:4.0).
Covers decision matrix for stuffing vs RAG, our stack constraints,
context budgeting, progressive loading, and smart compression.
2026-04-15 16:38:07 +00:00
10fd467b28 Merge pull request 'fix: resolve v2 harness import collision with explicit path loading (#716)' (#748) from burn/716-1776264183 into main 2026-04-15 16:04:04 +00:00
ba2d365669 fix: resolve v2 harness import collision with explicit path loading (closes #716)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 18s
2026-04-15 11:46:37 -04:00
Alexander Whitestone
b5386d45f4 fix: fix(ci): smoke workflow JSON parse + pytest coverage (fixes #715) (closes #746)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 13s
2026-04-15 11:38:07 -04:00
8900f22ddc fix(ci): JSON parse file-by-file, add pytest step
Some checks failed
Smoke Test / smoke (pull_request) Failing after 15s
- JSON validation: single Python process iterates all .json files
  (was: xargs piping all files to json.tool at once — fails)
- YAML validation: file-by-file for consistency
- Added pytest -q tests step (166 tests pass locally)
- Added set -euo pipefail for strict error handling
- Excluded .git/ from all find patterns

Fixes #715
2026-04-15 15:23:31 +00:00
Alexander Whitestone
a2115398d4 fix: smoke workflow JSON parse + add pytest step (closes #742) (closes #743)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 10m8s
2026-04-15 11:11:25 -04:00
475a64b167 fix: smoke workflow JSON parse + add pytest step (closes #742)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 10m26s
2026-04-15 15:07:55 +00:00
b7077a3c7e fix: smoke workflow JSON parse + add pytest step (#715)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 8m44s
2026-04-15 14:49:55 +00:00
5a696c184e Merge pull request 'feat: add NH Broadband install packet scaffold (closes #740)' (#741) from sprint/issue-740 into main 2026-04-15 11:57:34 +00:00
Alexander Whitestone
90d8daedcf feat: add NH Broadband install packet scaffold (closes #740)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 19s
2026-04-15 07:33:01 -04:00
Alexander Whitestone
35dad6211a fix: add Evennia VPS repair script (#534)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 23s
2026-04-15 01:50:26 -04:00
Alexander Whitestone
7addedda1c docs: add LAB-006 septic research packet (#531)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 23s
2026-04-15 01:50:06 -04:00
Alexander Whitestone
1050812bb5 fix: make big brain provider wiring generic (#543)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 31s
2026-04-15 01:26:14 -04:00
Alexander Whitestone
07e087a679 feat: add Bezalel Tailscale bootstrap scaffold (#535)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 18s
2026-04-15 01:25:37 -04:00
Alexander Whitestone
2946f9df73 feat: add Bezalel Gemma4 wiring scaffold (#544)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 26s
2026-04-15 01:17:32 -04:00
Timmy
231556e9ed fix: #648
Some checks failed
Smoke Test / smoke (pull_request) Failing after 22s
2026-04-15 01:09:16 -04:00
Alexander Whitestone
5d49b38ce3 feat: add fleet progression evaluator (#547)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 18s
2026-04-15 01:06:18 -04:00
Alexander Whitestone
d63654da22 ci: add self-healing smoke workflow (#549)
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Successful in 21s
Smoke Test / smoke (pull_request) Failing after 24s
2026-04-15 01:05:04 -04:00
Alexander Whitestone
c46caefed5 feat: prepare LAB-007 utility estimate packet (#532)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 25s
2026-04-15 01:03:12 -04:00
Alexander Whitestone
30e1fa19fa feat: add Gitea task delegation scaffold (#550)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 17s
2026-04-15 00:56:26 -04:00
Timmy
25dd988cc7 fix: #681
Some checks failed
Smoke Test / smoke (pull_request) Failing after 23s
2026-04-15 00:56:11 -04:00
Alexander Whitestone
0b4b20f62e feat: add fleet dispatch planning scaffold (#552)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 18s
2026-04-15 00:48:41 -04:00
Alexander Whitestone
8758f4e9d8 docs: add phase-4 sovereignty audit (#551)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 16s
2026-04-15 00:48:26 -04:00
Alexander Whitestone
b3359e1bae feat: add Bezalel Evennia world scaffold (#536)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 19s
2026-04-15 00:46:03 -04:00
Alexander Whitestone
cb46d56147 feat: add autonomous incident creation scaffold (#553)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 22s
2026-04-15 00:40:14 -04:00
Alexander Whitestone
cd7cb7bdc6 docs: finalize MemPalace evaluation report (#568)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 23s
2026-04-15 00:37:43 -04:00
Alexander Whitestone
12ec1af29f feat: ground unreachable horizon in a report (#545)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 30s
2026-04-15 00:33:46 -04:00
Timmy
9312e4dbee fix: #562
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 28s
Smoke Test / smoke (pull_request) Failing after 11m8s
Agent PR Gate / report (pull_request) Has been cancelled
2026-04-15 00:31:06 -04:00
Alexander Whitestone
173ce54eed feat: add timmy-home genome analysis (#670)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 19s
2026-04-15 00:28:53 -04:00
Alexander Whitestone
8d9e7cbf7e docs: record hermes-agent test finding (#668)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 18s
2026-04-15 00:26:33 -04:00
Alexander Whitestone
5186ab583b feat: add the-playground codebase genome (#671)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 21s
2026-04-15 00:25:18 -04:00
Alexander Whitestone
b90a15baca feat: codify phase-1 survival state (#548)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 11s
2026-04-15 00:24:25 -04:00
Alexander Whitestone
85bc612100 test: validate hermes-agent genome (#668) 2026-04-15 00:24:01 -04:00
Alexander Whitestone
9e120888c0 docs: add hermes-agent genome draft (#668) 2026-04-15 00:21:39 -04:00
Alexander Whitestone
ef5e0ec439 feat: add evennia mind palace architecture (refs #567)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 11s
2026-04-15 00:20:31 -04:00
Alexander Whitestone
9f55394639 test: define evennia mind palace acceptance for #567 2026-04-15 00:15:27 -04:00
Alexander Whitestone
6416b776db feat: add codebase genome pipeline (#665)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 27s
2026-04-15 00:14:55 -04:00
Alexander Whitestone
0e103dc8b7 feat: add the-door codebase genome (#673)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 8s
2026-04-15 00:09:42 -04:00
Alexander Whitestone
ae38b9b2bf docs: add timmy-config genome analysis (refs #669)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 17s
2026-04-15 00:05:15 -04:00
Alexander Whitestone
b334139fb5 feat: add fleet secret rotation playbook (#694)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 15s
2026-04-14 23:59:54 -04:00
Alexander Whitestone
6bbf6c4e0e ci: fix smoke JSON parsing for backup tests (#693)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 10m42s
2026-04-14 23:59:46 -04:00
Alexander Whitestone
1fed477af6 feat: sovereign DNS record management (#692)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 19s
2026-04-14 23:59:30 -04:00
Alexander Whitestone
6fbdbcf1c1 docs: add Hermes backup runbook (#693) 2026-04-14 23:57:41 -04:00
Alexander Whitestone
f8a9bae8fb feat: add encrypted Hermes backup pipeline (#693) 2026-04-14 23:56:46 -04:00
Alexander Whitestone
dda1e71029 wip: add the-testament genome analysis for #675
Some checks failed
Smoke Test / smoke (pull_request) Failing after 16s
2026-04-14 23:54:45 -04:00
5cc7b9b5a7 docs: QA triage action plan for #691
Some checks failed
Smoke Test / smoke (pull_request) Failing after 21s
Structured action plan converting cross-repo QA findings into
executable steps with owners, priorities, and verification.

Key findings addressed:
- P0: Production surfaces down (DNS/nginx), playground broken
- P1: 166 open PRs across 5 repos, 58 issues with duplicates
- P2: the-door crisis features blocked, no branch protection
- P3: Burn dedup gate, nightly triage cron

Priority order:
1. Fix DNS/nginx (crisis intervention reachable)
2. Close duplicate PRs (clear noise)
3. Review the-door PRs (mission-critical)
4. Fix the-playground (user-facing)
5. Enable branch protection
6. Build dedup gate
7. Nightly triage cron

Closes #691.
2026-04-14 23:51:40 -04:00
3b430114be feat: Codebase Genome for wolf (#683)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 20s
Complete GENOME.md for wolf (multi-model evaluation engine):
- Project overview and architecture diagram (Mermaid)
- Entry points and data flow (evaluation + task pipelines)
- Key abstractions (15 classes documented)
- API surface (CLI, providers, Gitea client)
- Scoring methodology (relevance/coherence/safety weights)
- Test coverage analysis with identified gaps
- Security considerations
- Sovereignty assessment
- File index with LOC counts

Repo 16/16. Closes #683.
2026-04-14 23:43:30 -04:00
Alexander Whitestone
8d1f9ed375 feat(#667): codebase_genome.py — test stub generator for uncovered functions
Some checks failed
Smoke Test / smoke (pull_request) Failing after 20s
AST-based tool that scans Python files, extracts function/method signatures,
and generates pytest test stubs for functions without existing tests.

Usage:
  python3 codebase_genome.py /path/to/repo
  python3 codebase_genome.py /path/to/repo -o tests/test_genome_generated.py

Features:
- AST parsing (no imports required, handles syntax errors gracefully)
- Extracts: function name, args, return type, decorators, class context
- Detects existing tests to avoid duplicates
- Generates: basic test + edge case test per function
- Skips private/dunder methods
- Configurable limit (--limit N)

Generated 30 test stubs for timmy-home as proof of concept.
2026-04-14 23:39:13 -04:00
b10974ef0b fix: generate GENOME.md for timmy-academy (#678)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 16s
Full codebase genome analysis of timmy-academy:
- Mermaid architecture diagram
- Entry points and data flow
- Key abstractions (AuditedCharacter, Commands, World)
- API surface, test coverage, security

Closes #678
2026-04-15 03:31:22 +00:00
8d60b6c693 fix(#676): Add GENOME.md for compounding-intelligence
Some checks failed
Smoke Test / smoke (pull_request) Failing after 25s
Complete codebase genome:
- Project overview and three-pipeline architecture
- Mermaid architecture diagram
- Entry points and data flow
- Knowledge schema and confidence scoring
- Key abstractions
- Test coverage analysis with gaps
- Security considerations
- Dependencies and status
2026-04-15 03:25:20 +00:00
f7843ae87f feat: GENOME.md for evennia-local-world (#677)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 30s
Full codebase analysis:
- Architecture diagram (Mermaid)
- Entry points and data flow
- Key abstractions (6 typeclasses, commands, scripts)
- API surface
- Test coverage gaps (8 critical paths identified)
- Security considerations (6 risks)
- Integration points (Timmy AI, Hermes, Federation)

Closes #677
2026-04-15 03:18:39 +00:00
ac25f2f9d4 fix(#679): Codebase Genome for turboquant
Some checks failed
Smoke Test / smoke (pull_request) Failing after 18s
2026-04-15 03:11:49 +00:00
Timmy
edca963e00 feat(#662): backlog cleanup script — bulk close issues with merged PRs
Some checks failed
Smoke Test / smoke (pull_request) Failing after 26s
Automates the recommendation from #662: bulk close issues whose
acceptance criteria are met and PRs are merged.

Features:
- Scans open issues for linked merged PRs
- Skips issues with assignees or active labels (epic, in-progress)
- Dry-run by default, --close to execute
- Adds closing comment for audit trail

Usage:
  python scripts/backlog_cleanup.py --dry-run
  python scripts/backlog_cleanup.py --close --limit 50

Refs #662
2026-04-14 22:56:17 -04:00
6dfd990f3a feat(#667): codebase-genome test suite generator — fill coverage gaps
Some checks failed
Smoke Test / smoke (pull_request) Failing after 19s
Scans Python codebases, identifies functions/methods, generates
pytest test cases for uncovered code.

Features:
- AST-based function discovery (args, returns, raises, docstrings)
- Module grouping and smart imports
- Edge case generation (None args, empty strings)
- Dry-run mode for preview
- Max-tests limit to prevent bloat
- Auto-generated marker for human review

Usage:
  python scripts/codebase-genome.py <dir> --dry-run
  python scripts/codebase-genome.py <dir> -o tests/test_genome.py
  python scripts/codebase-genome.py <dir> --max-tests 50

Refs #667
2026-04-14 22:36:27 -04:00
4582653bb4 feat: Codebase Genome — the-nexus full analysis (#672)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 16s
Complete GENOME.md for the-nexus repository:

- Project overview (3D world, Three.js, portal architecture)
- Architecture diagram (Mermaid)
- Entry points and data flow
- Key abstractions (NEXUS, SpatialMemory, Portal System, GOFAI)
- API surface (internal + external)
- Dependencies (Three.js, Python WebSocket)
- Test coverage gaps (7 critical paths untested)
- Security considerations (WebSocket auth, localStorage)
- Technical debt (4082-line app.js, no TypeScript)
- Migration notes from CLAUDE.md

Closes #672
2026-04-14 22:33:42 -04:00
Timmy
3b273f1345 feat: Codebase Test Generator — Fill Coverage Gaps (#667)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 23s
Add scripts/codebase_test_generator.py:
- AST analysis: finds all functions/methods in source files
- Coverage gap detection: identifies functions without tests
- Test generation: creates pytest test cases for each gap
- Priority: high (has return/raises), medium (public), low (private)
- Auto-generated markers for human review

Stats: 161 source files, 1219 functions, 873 coverage gaps found
Generated 50 tests: 14 pass, 30 skip (import), 6 fail (wrong args)

Usage:
  python scripts/codebase_test_generator.py --source . --output tests/test_genome_generated.py
2026-04-14 22:09:33 -04:00
Alexander Whitestone
8992c951a3 fix: [OPS] timmy-home backlog reduced from 220 to 50 — triage cadence needed (closes #685)
Some checks failed
Smoke Test / smoke (pull_request) Failing after 13s
2026-04-14 21:33:03 -04:00
Alexander Whitestone
038f1ab7f4 feat: Big Brain quality benchmark — gemma3:1b vs gemma3:27b (issue #576)
Ran 4 benchmark tasks on local gemma3:1b model with full quality analysis.
Big Brain (gemma3:27b on RunPod L40S) pod was offline (HTTP 404) during
benchmark — documented honestly with re-run instructions.

Tasks benchmarked:
1. Python Gitea webhook parser with HMAC-SHA256 verification
2. Evennia MUD framework architecture explanation
3. Fleet burn-down cron script for RunPod pods
4. Python async bug diagnosis and fix

Key finding: 1B model fails all tasks with hallucinated APIs, wrong
security primitives, fabricated technical details, and incorrect bug
diagnosis. Quality gap to 27B expected to be substantial.

Deliverable: timmy-config/docs/big-brain-benchmark.md
2026-04-13 20:49:02 -04:00
190 changed files with 25082 additions and 536 deletions

View File

@@ -0,0 +1,97 @@
name: Agent PR Gate
'on':
pull_request:
branches: [main]
jobs:
gate:
runs-on: ubuntu-latest
outputs:
syntax_status: ${{ steps.syntax.outcome }}
tests_status: ${{ steps.tests.outcome }}
criteria_status: ${{ steps.criteria.outcome }}
risk_level: ${{ steps.risk.outputs.level }}
steps:
- uses: actions/checkout@v4
with:
fetch-depth: 0
- uses: actions/setup-python@v5
with:
python-version: '3.11'
- name: Install CI dependencies
run: |
python3 -m pip install --quiet pyyaml pytest
- id: risk
name: Classify PR risk
run: |
BASE_REF="${GITHUB_BASE_REF:-main}"
git fetch origin "$BASE_REF" --depth 1
git diff --name-only "origin/$BASE_REF"...HEAD > /tmp/changed_files.txt
python3 scripts/agent_pr_gate.py classify-risk --files-file /tmp/changed_files.txt > /tmp/risk.json
python3 - <<'PY'
import json, os
with open('/tmp/risk.json', 'r', encoding='utf-8') as fh:
data = json.load(fh)
with open(os.environ['GITHUB_OUTPUT'], 'a', encoding='utf-8') as fh:
fh.write('level=' + data['risk'] + '\n')
PY
- id: syntax
name: Syntax and parse checks
continue-on-error: true
run: |
find . \( -name '*.yml' -o -name '*.yaml' \) | grep -v .gitea | xargs -r python3 -c "import sys,yaml; [yaml.safe_load(open(f)) for f in sys.argv[1:]]"
find . -name '*.json' | while read f; do python3 -m json.tool "$f" > /dev/null || exit 1; done
find . -name '*.py' | xargs -r python3 -m py_compile
find . -name '*.sh' | xargs -r bash -n
- id: tests
name: Test suite
continue-on-error: true
run: |
pytest -q --ignore=uni-wizard/v2/tests/test_author_whitelist.py
- id: criteria
name: PR criteria verification
continue-on-error: true
run: |
python3 scripts/agent_pr_gate.py validate-pr --event-path "$GITHUB_EVENT_PATH"
- name: Fail gate if any required check failed
if: steps.syntax.outcome != 'success' || steps.tests.outcome != 'success' || steps.criteria.outcome != 'success'
run: exit 1
report:
needs: gate
if: always()
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v4
- uses: actions/setup-python@v5
with:
python-version: '3.11'
- name: Post PR gate report
env:
GITEA_TOKEN: ${{ github.token }}
run: |
python3 scripts/agent_pr_gate.py comment \
--event-path "$GITHUB_EVENT_PATH" \
--token "$GITEA_TOKEN" \
--syntax "${{ needs.gate.outputs.syntax_status }}" \
--tests "${{ needs.gate.outputs.tests_status }}" \
--criteria "${{ needs.gate.outputs.criteria_status }}" \
--risk "${{ needs.gate.outputs.risk_level }}"
- name: Auto-merge low-risk clean PRs
if: needs.gate.result == 'success' && needs.gate.outputs.risk_level == 'low'
env:
GITEA_TOKEN: ${{ github.token }}
run: |
python3 scripts/agent_pr_gate.py merge \
--event-path "$GITHUB_EVENT_PATH" \
--token "$GITEA_TOKEN"

View File

@@ -0,0 +1,34 @@
name: Self-Healing Smoke
on:
pull_request:
push:
branches: [main]
jobs:
self-healing-smoke:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v4
- uses: actions/setup-python@v5
with:
python-version: '3.11'
- name: Shell syntax checks
run: |
bash -n scripts/fleet_health_probe.sh
bash -n scripts/auto_restart_agent.sh
bash -n scripts/backup_pipeline.sh
- name: Python compile checks
run: |
python3 -m py_compile uni-wizard/daemons/health_daemon.py
python3 -m py_compile scripts/fleet_milestones.py
python3 -m py_compile scripts/sovereign_health_report.py
python3 -m py_compile tests/docs/test_self_healing_infrastructure.py
python3 -m py_compile tests/docs/test_self_healing_ci.py
- name: Phase-2 doc tests
run: |
pytest -q tests/docs/test_self_healing_infrastructure.py tests/docs/test_self_healing_ci.py

View File

@@ -1,5 +1,5 @@
name: Smoke Test
on:
'on':
pull_request:
push:
branches: [main]
@@ -11,14 +11,30 @@ jobs:
- uses: actions/setup-python@v5
with:
python-version: '3.11'
- name: Parse check
- name: Install dependencies
run: |
find . -name '*.yml' -o -name '*.yaml' | grep -v .gitea | xargs -r python3 -c "import sys,yaml; [yaml.safe_load(open(f)) for f in sys.argv[1:]]"
find . -name '*.json' | xargs -r python3 -m json.tool > /dev/null
find . -name '*.py' | xargs -r python3 -m py_compile
find . -name '*.sh' | xargs -r bash -n
echo "PASS: All files parse"
python3 -m pip install --quiet pyyaml pytest
- name: YAML parse
run: |
find . \( -name '*.yml' -o -name '*.yaml' \) | grep -v .gitea | while read f; do python3 -c "import yaml; yaml.safe_load(open('$f'))" || { echo "FAIL: $f"; exit 1; }; done
echo "PASS: YAML files valid"
- name: JSON parse
run: |
find . -name '*.json' | while read f; do python3 -m json.tool "$f" > /dev/null || { echo "FAIL: $f"; exit 1; }; done
echo "PASS: JSON files valid"
- name: Python parse
run: |
find . -name '*.py' | while read f; do python3 -m py_compile "$f" || { echo "FAIL: $f"; exit 1; }; done
echo "PASS: Python files valid"
- name: Shell parse
run: |
find . -name '*.sh' | while read f; do bash -n "$f" || { echo "FAIL: $f"; exit 1; }; done
echo "PASS: Shell files valid"
- name: Secret scan
run: |
if grep -rE 'sk-or-|sk-ant-|ghp_|AKIA' . --include='*.yml' --include='*.py' --include='*.sh' 2>/dev/null | grep -v '.gitea' | grep -v 'detect_secrets' | grep -v 'test_trajectory_sanitize'; then exit 1; fi
echo "PASS: No secrets"
- name: Pytest
run: |
python3 -m pytest tests/ -q --tb=short
echo "PASS: All tests passed"

238
GENOME-timmy-academy.md Normal file
View File

@@ -0,0 +1,238 @@
# GENOME.md — timmy-academy
*Auto-generated by Codebase Genome Pipeline. 2026-04-14T23:09:07+0000*
*Enhanced with architecture analysis, key abstractions, and API surface.*
## Quick Facts
| Metric | Value |
|--------|-------|
| Source files | 48 |
| Test files | 1 |
| Config files | 1 |
| Total lines | 5,353 |
| Last commit | 395c9f7 Merge PR 'Add @who command' (#7) into master (2026-04-13) |
| Branch | master |
| Test coverage | 0% (35 untested modules) |
## What This Is
Timmy Academy is an Evennia-based MUD (Multi-User Dungeon) — a persistent text world where AI agents convene, train, and practice crisis response. It runs on Bezalel VPS (167.99.126.228) with telnet on port 4000 and web client on port 4001.
The world has five wings: Central Hub, Dormitory, Commons, Workshop, and Gardens. Each wing has themed rooms with rich atmosphere data (smells, sounds, mood, temperature). Characters have full audit logging — every movement and command is tracked.
## Architecture
```mermaid
graph TB
subgraph "Connections"
TELNET[Telnet :4000]
WEB[Web Client :4001]
end
subgraph "Evennia Core"
SERVER[Evennia Server]
PORTAL[Evennia Portal]
end
subgraph "Typeclasses"
CHAR[Character]
AUDIT[AuditedCharacter]
ROOM[Room]
EXIT[Exit]
OBJ[Object]
end
subgraph "Commands"
CMD_EXAM[CmdExamine]
CMD_ROOMS[CmdRooms]
CMD_STATUS[CmdStatus]
CMD_MAP[CmdMap]
CMD_ACADEMY[CmdAcademy]
CMD_SMELL[CmdSmell]
CMD_LISTEN[CmdListen]
CMD_WHO[CmdWho]
end
subgraph "World - Wings"
HUB[Central Hub]
DORM[Dormitory Wing]
COMMONS[Commons Wing]
WORKSHOP[Workshop Wing]
GARDENS[Gardens Wing]
end
subgraph "Hermes Bridge"
HERMES_CFG[hermes-agent/config.yaml]
BRIDGE[Agent Bridge]
end
TELNET --> SERVER
WEB --> PORTAL
PORTAL --> SERVER
SERVER --> CHAR
SERVER --> AUDIT
SERVER --> ROOM
SERVER --> EXIT
CHAR --> CMD_EXAM
CHAR --> CMD_STATUS
CHAR --> CMD_WHO
ROOM --> HUB
ROOM --> DORM
ROOM --> COMMONS
ROOM --> WORKSHOP
ROOM --> GARDENS
HERMES_CFG --> BRIDGE
BRIDGE --> SERVER
```
## Entry Points
| File | Purpose |
|------|---------|
| `server/conf/settings.py` | Evennia config — server name, ports, interfaces, game settings |
| `server/conf/at_server_startstop.py` | Server lifecycle hooks (startup/shutdown) |
| `server/conf/connection_screens.py` | Login/connection screen text |
| `commands/default_cmdsets.py` | Registers all custom commands with Evennia |
| `world/rebuild_world.py` | Rebuilds all rooms from source |
| `world/build_academy.ev` | Evennia batch script for initial world setup |
## Data Flow
```
Player connects (telnet/web)
-> Evennia Portal accepts connection
-> Server authenticates (Account typeclass)
-> Player puppets a Character
-> Character enters world (Room typeclass)
-> Commands processed through Command typeclass
-> AuditedCharacter logs every action
-> World responds with rich text + atmosphere data
```
## Key Abstractions
### Typeclasses (the world model)
| Class | File | Purpose |
|-------|------|---------|
| `Character` | `typeclasses/characters.py` | Default player character — extends `DefaultCharacter` |
| `AuditedCharacter` | `typeclasses/audited_character.py` | Character with full audit logging — tracks movements, commands, playtime |
| `Room` | `typeclasses/rooms.py` | Default room container |
| `Exit` | `typeclasses/exits.py` | Connections between rooms |
| `Object` | `typeclasses/objects.py` | Base object with `ObjectParent` mixin |
| `Account` | `typeclasses/accounts.py` | Player account (login identity) |
| `Channel` | `typeclasses/channels.py` | In-game communication channels |
| `Script` | `typeclasses/scripts.py` | Background/timed processes |
### AuditedCharacter — the flagship typeclass
The `AuditedCharacter` is the most important abstraction. It wraps every player action in logging:
- `at_pre_move()` — logs departure from current room
- `at_post_move()` — records arrival with timestamp and coordinates
- `at_pre_cmd()` — increments command counter, logs command + args
- `at_pre_puppet()` — starts session timer
- `at_post_unpuppet()` — calculates session duration, updates total playtime
- `get_audit_summary()` — returns JSON summary of all tracked metrics
Audit trail keeps last 1000 movements in `db.location_history`. Sensitive commands (password) are excluded from logging.
### Commands (the player interface)
| Command | Aliases | Purpose |
|---------|---------|---------|
| `examine` | `ex`, `exam` | Inspect room or object — shows description, atmosphere, objects, contents |
| `rooms` | — | List all rooms with wing color coding |
| `@status` | `status` | Show agent status: location, wing, mood, online players, uptime |
| `@map` | `map` | ASCII map of current wing |
| `@academy` | `academy` | Full academy overview with room counts |
| `smell` | `sniff` | Perceive room through atmosphere scent data |
| `listen` | `hear` | Perceive room through atmosphere sound data |
| `@who` | `who` | Show connected players with locations and idle time |
### World Structure (5 wings, 21+ rooms)
**Central Hub (LIMBO)** — Nexus connecting all wings. North=Dormitory, South=Workshop, East=Commons, West=Gardens.
**Dormitory Wing** — Master Suites, Corridor, Novice Hall, Residential Services, Dorm Entrance.
**Commons Wing** — Grand Commons Hall (main gathering, 60ft ceilings, marble columns), Hearthside Dining, Entertainment Gallery, Scholar's Corner, Upper Balcony.
**Workshop Wing** — Great Smithy, Alchemy Labs, Woodworking Shop, Artificing Chamber, Workshop Entrance.
**Gardens Wing** — Enchanted Grove, Herb Gardens, Greenhouse, Sacred Grove, Gardens Entrance.
Each room has rich `db.atmosphere` data: mood, lighting, sounds, smells, temperature.
## API Surface
### Web API
- `web/api/__init__.py` — Evennia REST API (Django REST Framework)
- `web/urls.py` — URL routing for web interface
- `web/admin/` — Django admin interface
- `web/website/` — Web frontend
### Telnet
- Standard MUD protocol on port 4000
- Supports MCCP (compression), MSDP (data), GMCP (protocol)
### Hermes Bridge
- `hermes-agent/config.yaml` — Configuration for AI agent connection
- Allows Hermes agents to connect as characters and interact with the world
## Dependencies
No `requirements.txt` or `pyproject.toml` found. Dependencies come from Evennia:
- **evennia** — MUD framework (Django-based)
- **django** — Web framework (via Evennia)
- **twisted** — Async networking (via Evennia)
## Test Coverage Analysis
| Metric | Value |
|--------|-------|
| Source modules | 35 |
| Test modules | 1 |
| Estimated coverage | 0% |
| Untested modules | 35 |
Only one test file exists: `tests/stress_test.py`. All 35 source modules are untested.
### Critical Untested Paths
1. **AuditedCharacter** — audit logging is the primary value-add. No tests verify movement tracking, command counting, or playtime calculation.
2. **Commands** — no tests for any of the 8 commands. The `@map` wing detection, `@who` session tracking, and atmosphere-based commands (`smell`, `listen`) are all untested.
3. **World rebuild**`rebuild_world.py` and `fix_world.py` can destroy and recreate the entire world. No tests ensure they produce valid output.
4. **Typeclass hooks**`at_pre_move`, `at_post_move`, `at_pre_cmd` etc. are never tested in isolation.
## Security Considerations
- ⚠️ Uses `eval()`/`exec()` — Evennia's inlinefuncs module uses eval for dynamic command evaluation. Risk level: inherent to MUD framework.
- ⚠️ References secrets/passwords — `settings.py` references `secret_settings.py` for sensitive config. Ensure this file is not committed.
- ⚠️ Telnet on 0.0.0.0 — server accepts connections from any IP. Consider firewall rules.
- ⚠️ Web client on 0.0.0.0 — same exposure as telnet. Ensure authentication is enforced.
- ⚠️ Agent bridge (`hermes-agent/config.yaml`) — verify credentials are not hardcoded.
## Configuration Files
- `server/conf/settings.py` — Main Evennia settings (server name, ports, typeclass paths)
- `hermes-agent/config.yaml` — Hermes agent bridge configuration
- `world/build_academy.ev` — Evennia batch build script
- `world/batch_cmds.ev` — Batch command definitions
## What's Missing
1. **Tests** — 0% coverage is a critical gap. Priority: AuditedCharacter hooks, command func() methods, world rebuild integrity.
2. **CI/CD** — No automated testing pipeline. No GitHub Actions or Gitea workflows.
3. **Documentation**`world/BUILDER_GUIDE.md` exists but no developer onboarding docs.
4. **Monitoring** — No health checks, no metrics export, no alerting on server crashes.
5. **Backup** — No automated database backup for the Evennia SQLite/PostgreSQL database.
---
*Generated by Codebase Genome Pipeline. Review and update manually.*

209
GENOME.md Normal file
View File

@@ -0,0 +1,209 @@
# GENOME.md — the-nexus
## Project Overview
`the-nexus` is a hybrid repo that combines three layers in one codebase:
1. A browser-facing world shell rooted in `index.html`, `boot.js`, `bootstrap.mjs`, `app.js`, `style.css`, `portals.json`, `vision.json`, `manifest.json`, and `gofai_worker.js`
2. A Python realtime bridge centered on `server.py` plus harness code under `nexus/`
3. A memory / fleet / operator layer spanning `mempalace/`, `mcp_servers/`, `multi_user_bridge.py`, and supporting scripts
The repo is not a clean single-purpose frontend and not just a backend harness. It is a mixed world/runtime/ops repository where browser rendering, WebSocket telemetry, MCP-driven game harnesses, and fleet memory tooling coexist.
Grounded repo facts from this checkout:
- Browser shell files exist at repo root: `index.html`, `app.js`, `style.css`, `manifest.json`, `gofai_worker.js`
- Data/config files also live at repo root: `portals.json`, `vision.json`
- Realtime bridge exists in `server.py`
- Game harnesses exist in `nexus/morrowind_harness.py` and `nexus/bannerlord_harness.py`
- Memory/fleet sync exists in `mempalace/tunnel_sync.py`
- Desktop/game automation MCP servers exist in `mcp_servers/desktop_control_server.py` and `mcp_servers/steam_info_server.py`
- Validation exists in `tests/test_browser_smoke.py`, `tests/test_portals_json.py`, `tests/test_index_html_integrity.py`, and `tests/test_repo_truth.py`
The current architecture is best understood as a sovereign world shell plus operator/game harness backend, with accumulated documentation drift from multiple restoration and migration efforts.
## Architecture Diagram
```mermaid
graph TD
browser[Index HTML Shell\nindex.html -> boot.js -> bootstrap.mjs -> app.js]
assets[Root Assets\nstyle.css\nmanifest.json\ngofai_worker.js]
data[World Data\nportals.json\nvision.json]
ws[Realtime Bridge\nserver.py\nWebSocket broadcast hub]
gofai[In-browser GOFAI\nSymbolicEngine\nNeuroSymbolicBridge\nsetupGOFAI/updateGOFAI]
harnesses[Python Harnesses\nnexus/morrowind_harness.py\nnexus/bannerlord_harness.py]
mcp[MCP Adapters\nmcp_servers/desktop_control_server.py\nmcp_servers/steam_info_server.py]
memory[Memory + Fleet\nmempalace/tunnel_sync.py\nmempalace.js]
bridge[Operator / MUD Bridge\nmulti_user_bridge.py\ncommands/timmy_commands.py]
tests[Verification\ntests/test_browser_smoke.py\ntests/test_portals_json.py\ntests/test_repo_truth.py]
docs[Contracts + Drift Docs\nBROWSER_CONTRACT.md\nREADME.md\nCLAUDE.md\nINVESTIGATION_ISSUE_1145.md]
browser --> assets
browser --> data
browser --> gofai
browser --> ws
harnesses --> mcp
harnesses --> ws
bridge --> ws
memory --> ws
tests --> browser
tests --> data
tests --> docs
docs --> browser
```
## Entry Points and Data Flow
### Primary entry points
- `index.html` — root browser entry point
- `boot.js` — startup selector; `tests/boot.test.js` shows it chooses file-mode vs HTTP/module-mode and injects `bootstrap.mjs` when served over HTTP
- `bootstrap.mjs` — module bootstrap for the browser shell
- `app.js` — main browser runtime; owns world state, GOFAI wiring, metrics polling, and portal/UI logic
- `server.py` — WebSocket broadcast bridge on `ws://0.0.0.0:8765`
- `nexus/morrowind_harness.py` — GamePortal/MCP harness for OpenMW Morrowind
- `nexus/bannerlord_harness.py` — GamePortal/MCP harness for Bannerlord
- `mempalace/tunnel_sync.py` — pulls remote fleet closets into the local palace over HTTP
- `multi_user_bridge.py` — HTTP bridge for multi-user chat/session integration
- `mcp_servers/desktop_control_server.py` — stdio MCP server exposing screenshots/mouse/keyboard control
### Data flow
1. Browser startup begins at `index.html`
2. `boot.js` decides whether the page is being served correctly; in HTTP mode it injects `bootstrap.mjs`
3. `bootstrap.mjs` hands off to `app.js`
4. `app.js` loads world configuration from `portals.json` and `vision.json`
5. `app.js` constructs the Three.js scene and in-browser reasoning components, including `SymbolicEngine`, `NeuroSymbolicBridge`, `setupGOFAI()`, and `updateGOFAI()`
6. Browser state and external runtimes connect through `server.py`, which broadcasts messages between connected clients
7. Python harnesses (`nexus/morrowind_harness.py`, `nexus/bannerlord_harness.py`) spawn MCP subprocesses for desktop control / Steam metadata, capture state, execute actions, and feed telemetry into the Nexus bridge
8. Memory/fleet tools like `mempalace/tunnel_sync.py` import remote palace data into local closets, extending what the operator/runtime layers can inspect
9. Tests validate both the static browser contract and the higher-level repo-truth/memory contracts
### Important repo-specific runtime facts
- `portals.json` is a JSON array of portal/world/operator entries; examples in this checkout include `morrowind`, `bannerlord`, `workshop`, `archive`, `chapel`, and `courtyard`
- `server.py` is a plain broadcast hub: clients send messages, the server forwards them to other connected clients
- `nexus/morrowind_harness.py` and `nexus/bannerlord_harness.py` both implement a GamePortal pattern with MCP subprocess clients over stdio and WebSocket telemetry uplink
- `mempalace/tunnel_sync.py` is not speculative; it is a real client that discovers remote wings, searches remote rooms, and writes `.closet.json` payloads locally
## Key Abstractions
### Browser runtime
- `app.js`
- Defines in-browser reasoning/state machinery, including `class SymbolicEngine`, `class NeuroSymbolicBridge`, `setupGOFAI()`, and `updateGOFAI()`
- Couples rendering, local symbolic reasoning, metrics polling, and portal/UI logic in one very large root module
- `BROWSER_CONTRACT.md`
- Acts like an executable architecture contract for the browser surface
- Declares required files, DOM IDs, Three.js expectations, provenance rules, and WebSocket expectations
### Realtime bridge
- `server.py`
- Single hub abstraction: a WebSocket broadcast server maintaining a `clients` set and forwarding messages from one client to the others
- This is the seam between browser shell, harnesses, and external telemetry producers
### GamePortal harness layer
- `nexus/morrowind_harness.py`
- `nexus/bannerlord_harness.py`
- Both define MCP client wrappers, `GameState` / `ActionResult`-style data classes, and an Observe-Decide-Act telemetry loop
- The harnesses are symmetric enough to be understood as reusable portal adapters with game-specific context injected on top
### Memory / fleet layer
- `mempalace/tunnel_sync.py`
- Encodes the fleet-memory sync client contract: discover wings, pull broad room queries, write closet files, support dry-run
- `mempalace.js`
- Minimal browser/Electron bridge to MemPalace commands via `window.electronAPI.execPython(...)`
- Important because it shows a second memory integration surface distinct from the Python fleet sync path
### Operator / interaction bridge
- `multi_user_bridge.py`
- `commands/timmy_commands.py`
- These bridge user-facing conversations or MUD/Evennia interactions back into Timmy/Nexus services
## API Surface
### Browser / static surface
- `index.html` served over HTTP
- `boot.js` exports `bootPage()`; verified by `node --test tests/boot.test.js`
- Data APIs are file-based inside the repo: `portals.json`, `vision.json`, `manifest.json`
### Network/runtime surface
- `python3 server.py`
- Starts the WebSocket bridge on port `8765`
- `python3 l402_server.py`
- Local HTTP microservice for cost-estimate style responses
- `python3 multi_user_bridge.py`
- Multi-user HTTP/chat bridge
### Harness / operator CLI surfaces
- `python3 nexus/morrowind_harness.py`
- `python3 nexus/bannerlord_harness.py`
- `python3 mempalace/tunnel_sync.py --peer <url> [--dry-run] [--n N]`
- `python3 mcp_servers/desktop_control_server.py`
- `python3 mcp_servers/steam_info_server.py`
### Validation surface
- `python3 -m pytest tests/test_portals_json.py tests/test_index_html_integrity.py tests/test_repo_truth.py -q`
- `node --test tests/boot.test.js`
- `python3 -m py_compile server.py nexus/morrowind_harness.py nexus/bannerlord_harness.py mempalace/tunnel_sync.py mcp_servers/desktop_control_server.py`
- `tests/test_browser_smoke.py` defines the higher-cost Playwright smoke contract for the world shell
## Test Coverage Gaps
Strongly covered in this checkout:
- `tests/test_portals_json.py` validates `portals.json`
- `tests/test_index_html_integrity.py` checks merge-marker/DOM-integrity regressions in `index.html`
- `tests/boot.test.js` verifies `boot.js` startup behavior
- `tests/test_repo_truth.py` validates the repo-truth documents
- Multiple `tests/test_mempalace_*.py` files cover the palace layer
- `tests/test_bannerlord_harness.py` exists for the Bannerlord harness
Notable gaps or weak seams:
- `nexus/morrowind_harness.py` is large and operationally critical, but the generated baseline still flags it as a gap relative to its size/complexity
- `mcp_servers/desktop_control_server.py` exposes high-power automation but has no obvious dedicated test file in the root `tests/` suite
- `app.js` is the dominant browser runtime file and mixes rendering, GOFAI, metrics, and integration logic in one place; browser smoke exists, but there is limited unit-level decomposition around those subsystems
- `mempalace.js` appears minimally bridged and stale relative to the richer Python MemPalace layer
- `multi_user_bridge.py` is a large integration surface and should be treated as high regression risk even though it is central to operator/chat flow
## Security Considerations
- `server.py` binds `HOST = "0.0.0.0"`, exposing the broadcast bridge beyond localhost unless network controls limit it
- The WebSocket bridge is a broadcast hub without visible authentication in `server.py`; connected clients are trusted to send messages into the bus
- `mcp_servers/desktop_control_server.py` exposes mouse/keyboard/screenshot control through a stdio MCP server. In any non-local or poorly isolated runtime, this is a privileged automation surface
- `app.js` contains hardcoded local/network endpoints such as `http://localhost:${L402_PORT}/api/cost-estimate` and `http://localhost:8082/metrics`; these are convenient for local development but create environment drift and deployment assumptions
- `app.js` also embeds explicit endpoint/status references like `ws://143.198.27.163:8765`, which is operationally brittle and the kind of hardcoded location data that drifts across environments
- `mempalace.js` shells out through `window.electronAPI.execPython(...)`; this is powerful and useful, but it is a clear trust boundary between UI and host execution
- `INVESTIGATION_ISSUE_1145.md` documents an earlier integrity hazard: agents writing to `public/nexus/` instead of canonical root paths. That path confusion is both an operational and security concern because it makes provenance harder to reason about
## Runtime Truth and Docs Drift
The most important architecture finding in this repo is not a class or subsystem. It is a truth mismatch.
- README.md says current `main` does not ship a browser 3D world
- CLAUDE.md declares root `app.js` and `index.html` as canonical frontend paths
- tests and browser contract now assume the root frontend exists
All three statements are simultaneously present in this checkout.
Grounded evidence:
- `README.md` still says the repo does not contain an active root frontend such as `index.html`, `app.js`, or `style.css`
- the current checkout does contain `index.html`, `app.js`, `style.css`, `manifest.json`, and `gofai_worker.js`
- `BROWSER_CONTRACT.md` explicitly treats those root files as required browser assets
- `tests/test_browser_smoke.py` serves those exact files and validates DOM/WebGL contracts against them
- `tests/test_index_html_integrity.py` assumes `index.html` is canonical and production-relevant
- `CLAUDE.md` says frontend code lives at repo root and explicitly warns against `public/nexus/`
- `INVESTIGATION_ISSUE_1145.md` explains why `public/nexus/` is a bad/corrupt duplicate path and confirms the real classical AI code lives in root `app.js`
The honest conclusion:
- The repo contains a partially restored or actively re-materialized browser surface
- The docs are preserving an older migration truth while the runtime files and smoke contracts describe a newer present-tense truth
- Any future work in `the-nexus` must choose one truth and align `README.md`, `CLAUDE.md`, smoke tests, and file layout around it
That drift is itself a critical architectural fact and should be treated as first-order design debt, not a side note.

View File

@@ -0,0 +1,21 @@
fleet_rotation_backup_root: /var/lib/timmy/secret-rotations
fleet_secret_targets:
ezra:
env_file: /root/wizards/ezra/home/.env
ssh_authorized_keys_file: /root/.ssh/authorized_keys
services:
- hermes-ezra.service
- openclaw-ezra.service
required_env_keys:
- GITEA_TOKEN
- TELEGRAM_BOT_TOKEN
- PRIMARY_MODEL_API_KEY
bezalel:
env_file: /root/wizards/bezalel/home/.env
ssh_authorized_keys_file: /root/.ssh/authorized_keys
services:
- hermes-bezalel.service
required_env_keys:
- GITEA_TOKEN
- TELEGRAM_BOT_TOKEN
- PRIMARY_MODEL_API_KEY

View File

@@ -0,0 +1,79 @@
fleet_secret_bundle:
ezra:
env:
GITEA_TOKEN: !vault |
$ANSIBLE_VAULT;1.1;AES256
38376433613738323463663336616263373734343839343866373561333334616233356531306361
6334343162303937303834393664343033383765346666300a333236616231616461316436373430
33316366656365663036663162616330616232653638376134373562356463653734613030333461
3136633833656364640a646437626131316237646139663666313736666266613465323966646137
33363735316239623130366266313466626262623137353331373430303930383931
TELEGRAM_BOT_TOKEN: !vault |
$ANSIBLE_VAULT;1.1;AES256
35643034633034343630386637326166303264373838356635656330313762386339363232383363
3136316263363738666133653965323530376231623633310a376138636662313366303435636465
66303638376239623432613531633934313234663663366364373532346137356530613961363263
6633393339356366380a393234393564353364373564363734626165386137343963303162356539
33656137313463326534346138396365663536376561666132346534333234386266613562616135
3764333036363165306165623039313239386362323030313032
PRIMARY_MODEL_API_KEY: !vault |
$ANSIBLE_VAULT;1.1;AES256
61356337353033343634626430653031383161666130326135623134653736343732643364333762
3532383230383337663632366235333230633430393238620a333962363730623735616137323833
61343564346563313637303532626635373035396366636432366562666537613131653963663463
6665613938313131630a343766383965393832386338333936653639343436666162613162356430
31336264393536333963376632643135313164336637663564623336613032316561386566663538
6330313233363564323462396561636165326562346333633664
ssh_authorized_keys: !vault |
$ANSIBLE_VAULT;1.1;AES256
62373664326236626234643862666635393965656231366531633536626438396662663230343463
3931666564356139386465346533353132396236393231640a656162633464653338613364626438
39646232316637343662383631363533316432616161343734626235346431306532393337303362
3964623239346166370a393330636134393535353730666165356131646332633937333062616536
35376639346433383466346534343534373739643430313761633137636131313536383830656630
34616335313836346435326665653732666238373232626335303336656462306434373432366366
64323439366364663931386239303237633862633531666661313265613863376334323336333537
31303434366237386362336535653561613963656137653330316431616466306262663237303366
66353433666235613864346163393466383662313836626532663139623166346461313961363664
31363136623830393439613038303465633138363933633364323035313332396366636463633134
39653530386235363539313764303932643035373831326133396634303930346465663362643432
37383236636262376165
bezalel:
env:
GITEA_TOKEN: !vault |
$ANSIBLE_VAULT;1.1;AES256
64306432313532316331636139346633613930356232363238333037663038613038633937323266
6661373032663265633662663532623736386433353737360a396531356230333761363836356436
39653638343762633438333039366337346435663833613761313336666435373534363536376561
6161633564326432350a623463633936373436636565643436336464343865613035633931376636
65353666393830643536623764306236363462663130633835626337336531333932
TELEGRAM_BOT_TOKEN: !vault |
$ANSIBLE_VAULT;1.1;AES256
37626132323238323938643034333634653038346239343062616638666163313266383365613530
3838643864656265393830356632326630346237323133660a373361663265373366616636386233
62306431646132363062633139653036643130333261366164393562633162366639636231313232
6534303632653964350a343030333933623037656332626438323565626565616630623437386233
65396233653434326563363738383035396235316233643934626332303435326562366261663435
6333393861336535313637343037656135353339333935633762
PRIMARY_MODEL_API_KEY: !vault |
$ANSIBLE_VAULT;1.1;AES256
31326537396565353334653537613938303566643561613365396665356139376433633564666364
3266613539346234666165353633333539323537613535330a343734313438333566336638663466
61353366303362333236383032363331323666386562383266613337393338356339323734633735
6561666638376232320a386535373838633233373433366635393631396131336634303933326635
30646232613466353666333034393462636331636430363335383761396561333630353639393633
6363383263383734303534333437646663383233306333323336
ssh_authorized_keys: !vault |
$ANSIBLE_VAULT;1.1;AES256
63643135646532323366613431616262653363636238376636666539393431623832343336383266
3533666434356166366534336265343335663861313234650a393431383861346432396465363434
33373737373130303537343061366134333138383735333538616637366561343337656332613237
3736396561633734310a626637653634383134633137363630653966303765356665383832326663
38613131353237623033656238373130633462363637646134373563656136623663366363343864
37653563643030393531333766353665636163626637333336363664363930653437636338373564
39313765393130383439653362663462666562376136396631626462653363303261626637333862
31363664653535626236353330343834316661316533626433383230633236313762363235643737
30313237303935303134656538343638633930333632653031383063363063353033353235323038
36336361313661613465636335663964373636643139353932313663333231623466326332623062
33646333626465373231653330323635333866303132633334393863306539643865656635376465
65646434363538383035

View File

@@ -0,0 +1,3 @@
[fleet]
ezra ansible_host=143.198.27.163 ansible_user=root
bezalel ansible_host=67.205.155.108 ansible_user=root

View File

@@ -0,0 +1,185 @@
---
- name: Rotate vaulted fleet secrets
hosts: fleet
gather_facts: false
any_errors_fatal: true
serial: 100%
vars_files:
- ../inventory/group_vars/fleet_secrets.vault.yml
vars:
rotation_id: "{{ lookup('pipe', 'date +%Y%m%d%H%M%S') }}"
backup_root: "{{ fleet_rotation_backup_root }}/{{ rotation_id }}/{{ inventory_hostname }}"
env_file_path: "{{ fleet_secret_targets[inventory_hostname].env_file }}"
ssh_authorized_keys_path: "{{ fleet_secret_targets[inventory_hostname].ssh_authorized_keys_file }}"
env_backup_path: "{{ backup_root }}/env.before"
ssh_backup_path: "{{ backup_root }}/authorized_keys.before"
staged_env_path: "{{ backup_root }}/env.candidate"
staged_ssh_path: "{{ backup_root }}/authorized_keys.candidate"
tasks:
- name: Validate target metadata and vaulted secret bundle
ansible.builtin.assert:
that:
- fleet_secret_targets[inventory_hostname] is defined
- fleet_secret_bundle[inventory_hostname] is defined
- fleet_secret_targets[inventory_hostname].services | length > 0
- fleet_secret_targets[inventory_hostname].required_env_keys | length > 0
- fleet_secret_bundle[inventory_hostname].env is defined
- fleet_secret_bundle[inventory_hostname].ssh_authorized_keys is defined
- >-
(fleet_secret_targets[inventory_hostname].required_env_keys
| difference(fleet_secret_bundle[inventory_hostname].env.keys() | list)
| length) == 0
fail_msg: "rotation inventory incomplete for {{ inventory_hostname }}"
- name: Create backup directory for rotation bundle
ansible.builtin.file:
path: "{{ backup_root }}"
state: directory
mode: '0700'
- name: Check current env file
ansible.builtin.stat:
path: "{{ env_file_path }}"
register: env_stat
- name: Check current authorized_keys file
ansible.builtin.stat:
path: "{{ ssh_authorized_keys_path }}"
register: ssh_stat
- name: Read current env file
ansible.builtin.slurp:
src: "{{ env_file_path }}"
register: env_current
when: env_stat.stat.exists
- name: Read current authorized_keys file
ansible.builtin.slurp:
src: "{{ ssh_authorized_keys_path }}"
register: ssh_current
when: ssh_stat.stat.exists
- name: Save env rollback snapshot
ansible.builtin.copy:
content: "{{ env_current.content | b64decode }}"
dest: "{{ env_backup_path }}"
mode: '0600'
when: env_stat.stat.exists
- name: Save authorized_keys rollback snapshot
ansible.builtin.copy:
content: "{{ ssh_current.content | b64decode }}"
dest: "{{ ssh_backup_path }}"
mode: '0600'
when: ssh_stat.stat.exists
- name: Build staged env candidate
ansible.builtin.copy:
content: "{{ (env_current.content | b64decode) if env_stat.stat.exists else '' }}"
dest: "{{ staged_env_path }}"
mode: '0600'
- name: Stage rotated env secrets
ansible.builtin.lineinfile:
path: "{{ staged_env_path }}"
regexp: "^{{ item.key }}="
line: "{{ item.key }}={{ item.value }}"
create: true
loop: "{{ fleet_secret_bundle[inventory_hostname].env | dict2items }}"
loop_control:
label: "{{ item.key }}"
no_log: true
- name: Ensure SSH directory exists
ansible.builtin.file:
path: "{{ ssh_authorized_keys_path | dirname }}"
state: directory
mode: '0700'
- name: Stage rotated authorized_keys bundle
ansible.builtin.copy:
content: "{{ fleet_secret_bundle[inventory_hostname].ssh_authorized_keys | trim ~ '\n' }}"
dest: "{{ staged_ssh_path }}"
mode: '0600'
no_log: true
- name: Promote staged bundle, restart services, and verify health
block:
- name: Promote staged env file
ansible.builtin.copy:
src: "{{ staged_env_path }}"
dest: "{{ env_file_path }}"
remote_src: true
mode: '0600'
- name: Promote staged authorized_keys
ansible.builtin.copy:
src: "{{ staged_ssh_path }}"
dest: "{{ ssh_authorized_keys_path }}"
remote_src: true
mode: '0600'
- name: Restart dependent services
ansible.builtin.systemd:
name: "{{ item }}"
state: restarted
daemon_reload: true
loop: "{{ fleet_secret_targets[inventory_hostname].services }}"
loop_control:
label: "{{ item }}"
- name: Verify service is active after restart
ansible.builtin.command: "systemctl is-active {{ item }}"
register: service_status
changed_when: false
failed_when: service_status.stdout.strip() != 'active'
loop: "{{ fleet_secret_targets[inventory_hostname].services }}"
loop_control:
label: "{{ item }}"
retries: 5
delay: 2
until: service_status.stdout.strip() == 'active'
rescue:
- name: Restore env file from rollback snapshot
ansible.builtin.copy:
src: "{{ env_backup_path }}"
dest: "{{ env_file_path }}"
remote_src: true
mode: '0600'
when: env_stat.stat.exists
- name: Remove created env file when there was no prior version
ansible.builtin.file:
path: "{{ env_file_path }}"
state: absent
when: not env_stat.stat.exists
- name: Restore authorized_keys from rollback snapshot
ansible.builtin.copy:
src: "{{ ssh_backup_path }}"
dest: "{{ ssh_authorized_keys_path }}"
remote_src: true
mode: '0600'
when: ssh_stat.stat.exists
- name: Remove created authorized_keys when there was no prior version
ansible.builtin.file:
path: "{{ ssh_authorized_keys_path }}"
state: absent
when: not ssh_stat.stat.exists
- name: Restart services after rollback
ansible.builtin.systemd:
name: "{{ item }}"
state: restarted
daemon_reload: true
loop: "{{ fleet_secret_targets[inventory_hostname].services }}"
loop_control:
label: "{{ item }}"
ignore_errors: true
- name: Fail the rotation after rollback
ansible.builtin.fail:
msg: "Rotation failed for {{ inventory_hostname }}. Previous secrets restored from {{ backup_root }}."

275
codebase_genome.py Normal file
View File

@@ -0,0 +1,275 @@
#!/usr/bin/env python3
"""
codebase_genome.py — Analyze a repo and generate test stubs for uncovered functions.
Scans Python files, extracts function/class/method signatures via AST,
and generates pytest test cases with edge cases.
Usage:
python3 codebase_genome.py /path/to/repo
python3 codebase_genome.py /path/to/repo --output tests/test_genome_generated.py
"""
import ast
import os
import sys
import argparse
from pathlib import Path
class FunctionInfo:
def __init__(self, name, filepath, lineno, args, returns, decorators, is_method=False, class_name=None):
self.name = name
self.filepath = filepath
self.lineno = lineno
self.args = args # list of arg names
self.returns = returns # return annotation or None
self.decorators = decorators
self.is_method = is_method
self.class_name = class_name
@property
def qualified_name(self):
if self.class_name:
return f"{self.class_name}.{self.name}"
return self.name
@property
def import_path(self):
"""Module path for import (e.g., 'mymodule.sub.Class.method')."""
rel = Path(self.filepath).with_suffix('')
parts = list(rel.parts)
# Remove common prefixes
if parts and parts[0] in ('src', 'lib'):
parts = parts[1:]
module = '.'.join(parts)
if self.class_name:
return f"{module}.{self.class_name}.{self.name}"
return f"{module}.{self.name}"
@property
def module_path(self):
rel = Path(self.filepath).with_suffix('')
parts = list(rel.parts)
if parts and parts[0] in ('src', 'lib'):
parts = parts[1:]
return '.'.join(parts)
def extract_functions(filepath: str) -> list:
"""Extract all function definitions from a Python file via AST."""
try:
source = open(filepath).read()
tree = ast.parse(source, filename=filepath)
except (SyntaxError, UnicodeDecodeError):
return []
functions = []
class FuncVisitor(ast.NodeVisitor):
def __init__(self):
self.current_class = None
def visit_ClassDef(self, node):
old_class = self.current_class
self.current_class = node.name
self.generic_visit(node)
self.current_class = old_class
def visit_FunctionDef(self, node):
args = [a.arg for a in node.args.args]
if args and args[0] == 'self':
args = args[1:]
returns = None
if node.returns:
if isinstance(node.returns, ast.Name):
returns = node.returns.id
elif isinstance(node.returns, ast.Constant):
returns = str(node.returns.value)
decorators = []
for d in node.decorator_list:
if isinstance(d, ast.Name):
decorators.append(d.id)
elif isinstance(d, ast.Attribute):
decorators.append(d.attr)
functions.append(FunctionInfo(
name=node.name,
filepath=filepath,
lineno=node.lineno,
args=args,
returns=returns,
decorators=decorators,
is_method=self.current_class is not None,
class_name=self.current_class,
))
self.generic_visit(node)
visit_AsyncFunctionDef = visit_FunctionDef
visitor = FuncVisitor()
visitor.visit(tree)
return functions
def generate_test(func: FunctionInfo, existing_tests: set) -> str:
"""Generate a pytest test function for a given function."""
if func.name in existing_tests:
return ''
# Skip private/dunder methods
if func.name.startswith('_') and not func.name.startswith('__'):
return ''
if func.name.startswith('__') and func.name.endswith('__'):
return ''
lines = []
# Generate imports
module = func.module_path.replace('/', '.').lstrip('.')
if func.class_name:
lines.append(f"from {module} import {func.class_name}")
else:
lines.append(f"from {module} import {func.name}")
lines.append('')
lines.append('')
# Test function name
test_name = f"test_{func.qualified_name.replace('.', '_')}"
# Determine args for the test call
args_str = ', '.join(func.args)
lines.append(f"def {test_name}():")
lines.append(f' """Test {func.qualified_name} (line {func.lineno} in {func.filepath})."""')
if func.is_method:
lines.append(f" # TODO: instantiate {func.class_name} with valid args")
lines.append(f" obj = {func.class_name}()")
lines.append(f" result = obj.{func.name}({', '.join('None' for _ in func.args) if func.args else ''})")
else:
if func.args:
lines.append(f" # TODO: provide valid arguments for: {args_str}")
lines.append(f" result = {func.name}({', '.join('None' for _ in func.args)})")
else:
lines.append(f" result = {func.name}()")
lines.append(f" assert result is not None or result is None # TODO: real assertion")
lines.append('')
lines.append('')
# Edge cases
lines.append(f"def {test_name}_edge_cases():")
lines.append(f' """Edge cases for {func.qualified_name}."""')
if func.args:
lines.append(f" # Test with empty/zero/None args")
if func.is_method:
lines.append(f" obj = {func.class_name}()")
for arg in func.args:
lines.append(f" # obj.{func.name}({arg}=...) # TODO: test with invalid {arg}")
else:
for arg in func.args:
lines.append(f" # {func.name}({arg}=...) # TODO: test with invalid {arg}")
else:
lines.append(f" # {func.qualified_name} takes no args — test idempotency")
if func.is_method:
lines.append(f" obj = {func.class_name}()")
lines.append(f" r1 = obj.{func.name}()")
lines.append(f" r2 = obj.{func.name}()")
lines.append(f" # assert r1 == r2 # TODO: uncomment if deterministic")
else:
lines.append(f" r1 = {func.name}()")
lines.append(f" r2 = {func.name}()")
lines.append(f" # assert r1 == r2 # TODO: uncomment if deterministic")
lines.append('')
lines.append('')
return '\n'.join(lines)
def scan_repo(repo_path: str) -> list:
"""Scan all Python files in a repo and extract functions."""
all_functions = []
for root, dirs, files in os.walk(repo_path):
# Skip hidden dirs, __pycache__, .git, venv, node_modules
dirs[:] = [d for d in dirs if not d.startswith('.') and d not in ('__pycache__', 'venv', 'node_modules', 'env')]
for f in files:
if f.endswith('.py') and not f.startswith('_'):
filepath = os.path.join(root, f)
relpath = os.path.relpath(filepath, repo_path)
funcs = extract_functions(filepath)
# Update filepath to relative
for func in funcs:
func.filepath = relpath
all_functions.extend(funcs)
return all_functions
def find_existing_tests(repo_path: str) -> set:
"""Find function names that already have tests."""
tested = set()
tests_dir = os.path.join(repo_path, 'tests')
if not os.path.isdir(tests_dir):
return tested
for root, dirs, files in os.walk(tests_dir):
for f in files:
if f.startswith('test_') and f.endswith('.py'):
try:
source = open(os.path.join(root, f)).read()
tree = ast.parse(source)
for node in ast.walk(tree):
if isinstance(node, ast.FunctionDef) and node.name.startswith('test_'):
# Extract function name from test name
name = node.name[5:] # strip 'test_'
tested.add(name)
except (SyntaxError, UnicodeDecodeError):
pass
return tested
def main():
parser = argparse.ArgumentParser(description='Generate test stubs for uncovered functions')
parser.add_argument('repo', help='Path to repository')
parser.add_argument('--output', '-o', default=None, help='Output file (default: stdout)')
parser.add_argument('--limit', '-n', type=int, default=50, help='Max tests to generate')
args = parser.parse_args()
repo = os.path.abspath(args.repo)
if not os.path.isdir(repo):
print(f"Error: {repo} is not a directory", file=sys.stderr)
sys.exit(1)
functions = scan_repo(repo)
existing = find_existing_tests(repo)
# Filter to untested functions
untested = [f for f in functions if f.name not in existing and not f.name.startswith('_')]
print(f"Found {len(functions)} functions, {len(untested)} untested", file=sys.stderr)
# Generate tests
output = []
output.append('"""Auto-generated test stubs from codebase_genome.py.\n')
output.append('These are starting points — fill in real assertions and args.\n"""')
output.append('import pytest')
output.append('')
generated = 0
for func in untested[:args.limit]:
test = generate_test(func, set())
if test:
output.append(test)
generated += 1
content = '\n'.join(output)
if args.output:
with open(args.output, 'w') as f:
f.write(content)
print(f"Generated {generated} test stubs → {args.output}", file=sys.stderr)
else:
print(content)
if __name__ == '__main__':
main()

View File

@@ -0,0 +1,534 @@
# GENOME.md — compounding-intelligence
*Generated: 2026-04-21 07:23:18 UTC | Refreshed for timmy-home #676 from `Timmy_Foundation/compounding-intelligence` @ `fe8a70a` on `main`*
## Project Overview
`compounding-intelligence` is a Python-first analysis toolkit for turning prior agent work into reusable fleet knowledge.
At a high level it does four things:
1. reads Hermes session transcripts and diff/session artifacts
2. extracts durable knowledge into a structured store
3. assembles bootstrap context for future sessions
4. mines the corpus for higher-order opportunities: automation, refactors, performance, knowledge gaps, and issue-priority changes
The repo's own README still presents the system as three largely planned pipelines. That is now stale.
Current repo truth from live inspection:
- tracked files: 56
- 33 Python files
- 15 test Python files
- Python LOC: 8,394
- workflow files: `.gitea/workflows/test.yml`
- persistent data fixtures: 5 JSONL files under `test_sessions/`
- existing target-repo genome already present upstream: `GENOME.md`
Most important architecture fact:
- this repo is no longer just prompt scaffolding for a future harvester/bootstrapper/measurer loop
- it already contains a growing family of concrete analysis engines under `scripts/`
Largest Python modules by size:
- `scripts/priority_rebalancer.py` — 682 lines
- `scripts/automation_opportunity_finder.py` — 554 lines
- `scripts/perf_bottleneck_finder.py` — 551 lines
- `scripts/improvement_proposals.py` — 451 lines
- `scripts/harvester.py` — 447 lines
- `scripts/bootstrapper.py` — 359 lines
- `scripts/sampler.py` — 353 lines
- `scripts/dead_code_detector.py` — 282 lines
## Architecture
The repo is best understood as three layers: ingestion, knowledge storage/bootstrap, and meta-analysis.
```mermaid
flowchart TD
A[Hermes session JSONL] --> B[session_reader.py]
B --> C[harvester.py]
B --> D[session_pair_harvester.py]
C --> E[knowledge/index.json]
C --> F[knowledge/global/*.yaml or .md]
C --> G[knowledge/repos/*.yaml]
C --> H[knowledge/agents/*]
E --> I[bootstrapper.py]
F --> I
G --> I
H --> I
I --> J[Bootstrapped session context]
E --> K[knowledge_staleness_check.py]
E --> L[priority_rebalancer.py]
E --> M[improvement_proposals.py]
N[test_sessions/*.jsonl] --> C
N --> D
N --> M
O[repo source tree] --> P[knowledge_gap_identifier.py]
O --> Q[dead_code_detector.py]
O --> R[automation_opportunity_finder.py]
O --> S[perf_bottleneck_finder.py]
O --> T[dependency_graph.py]
O --> U[diff_analyzer.py]
O --> V[refactoring_opportunity_finder.py]
W[Gitea issues API] --> L
L --> X[metrics/priority_report.json]
L --> Y[metrics/priority_suggestions.md]
```
What exists today:
- transcript parsing: `scripts/session_reader.py`
- knowledge extraction + dedup + writing: `scripts/harvester.py`
- context assembly: `scripts/bootstrapper.py`
- pair harvesting: `scripts/session_pair_harvester.py`
- staleness detection: `scripts/knowledge_staleness_check.py`
- gap analysis: `scripts/knowledge_gap_identifier.py`
- improvement mining: `scripts/improvement_proposals.py`
- automation mining: `scripts/automation_opportunity_finder.py`
- priority scoring against Gitea: `scripts/priority_rebalancer.py`
- diff scanning: `scripts/diff_analyzer.py`
- dead code analysis: `scripts/dead_code_detector.py`
What exists but is currently broken or incomplete:
- `scripts/refactoring_opportunity_finder.py` is still a stub that only emits sample proposals
- `scripts/perf_bottleneck_finder.py` does not parse
- `scripts/dependency_graph.py` does not parse
## Runtime Truth and Docs Drift
The repo ships its own `GENOME.md`, but that document is materially stale relative to the current codebase.
The strongest drift example:
- upstream `GENOME.md` says core pipeline scripts such as `harvester.py`, `bootstrapper.py`, `measurer.py`, and `session_reader.py` are planned or not yet implemented
- live source inspection shows `scripts/harvester.py`, `scripts/bootstrapper.py`, and `scripts/session_reader.py` are real, non-trivial implementations
- live source inspection also shows additional implemented engines not foregrounded by the README's original three-pipeline framing:
- `scripts/priority_rebalancer.py`
- `scripts/automation_opportunity_finder.py`
- `scripts/improvement_proposals.py`
- `scripts/knowledge_gap_identifier.py`
- `scripts/dead_code_detector.py`
- `scripts/session_pair_harvester.py`
- `scripts/diff_analyzer.py`
So the honest current description is:
- README = founding vision
- existing target-repo `GENOME.md` = partially outdated snapshot
- source + tests = current system truth
This is not a repo with only a single harvester/bootstrapper loop anymore. It is becoming a general-purpose compounding-analysis workbench.
## Entry Points
### 1. CI / canonical test entry point
The only checked-in workflow is `.gitea/workflows/test.yml`.
It installs:
- `requirements.txt`
Then runs:
```bash
make test
```
The Makefile defines:
```make
python3 -m pytest tests/test_ci_config.py scripts/test_*.py -v
```
This is the repo's canonical automation contract today.
### 2. Knowledge extraction entry point
`scripts/harvester.py`
Docstring usage:
```bash
python3 harvester.py --session ~/.hermes/sessions/session_xxx.jsonl --output knowledge/
python3 harvester.py --batch --since 2026-04-01 --limit 100
python3 harvester.py --session session.jsonl --dry-run
```
This is the main LLM-integrated path.
### 3. Session bootstrap entry point
`scripts/bootstrapper.py`
Docstring usage:
```bash
python3 bootstrapper.py --repo the-nexus --agent mimo-sprint
python3 bootstrapper.py --repo timmy-home --global
python3 bootstrapper.py --global
python3 bootstrapper.py --repo the-nexus --max-tokens 1000
```
### 4. Priority rebalancer entry point
`scripts/priority_rebalancer.py`
Docstring usage:
```bash
python3 scripts/priority_rebalancer.py --org Timmy_Foundation
python3 scripts/priority_rebalancer.py --org Timmy_Foundation --repo compounding-intelligence
python3 scripts/priority_rebalancer.py --org Timmy_Foundation --dry-run
python3 scripts/priority_rebalancer.py --org Timmy_Foundation --apply
```
### 5. Secondary analysis engines
Additional operational entry points exist in `scripts/`:
- `automation_opportunity_finder.py`
- `improvement_proposals.py`
- `knowledge_gap_identifier.py`
- `knowledge_staleness_check.py`
- `dead_code_detector.py`
- `diff_analyzer.py`
- `sampler.py`
- `gitea_issue_parser.py`
- `session_pair_harvester.py`
### 6. Seed knowledge content
The knowledge store is not empty scaffolding.
Concrete checked-in knowledge already exists at:
- `knowledge/repos/hermes-agent.yaml`
- `knowledge/repos/the-nexus.yaml`
- `knowledge/global/pitfalls.yaml`
- `knowledge/global/tool-quirks.yaml`
- `knowledge/index.json`
- `knowledge/SCHEMA.md`
## Data Flow
### Flow A — transcript to durable knowledge
1. Raw session JSONL enters via `scripts/session_reader.py`.
2. `read_session()` loads the transcript.
3. `extract_conversation()` strips to meaningful user/assistant/system turns.
4. `truncate_for_context()` compresses long sessions to head + tail.
5. `messages_to_text()` converts structured turns to a plain-text transcript block.
6. `scripts/harvester.py` loads `templates/harvest-prompt.md`.
7. The harvester calls an LLM endpoint, parses the JSON response, validates facts, fingerprints them, deduplicates, then writes `knowledge/index.json` and human-readable per-domain files.
### Flow B — durable knowledge to session bootstrap
1. `scripts/bootstrapper.py` loads `knowledge/index.json`.
2. It filters facts by repo, agent, and global scope.
3. It sorts them by confidence and category priority.
4. It optionally merges markdown knowledge from repo-specific, agent-specific, and global files.
5. It truncates the result to a token budget and emits a bootstrap context block.
### Flow C — corpus to meta-analysis
Several scripts mine the repo and/or session corpus for second-order leverage:
- `scripts/improvement_proposals.py` mines repeated errors, slow tools, manual processes, and retries into proposal objects
- `scripts/automation_opportunity_finder.py` scans transcripts, scripts, docs, and cron jobs for automatable work
- `scripts/knowledge_gap_identifier.py` cross-references code, docs, and tests
- `scripts/priority_rebalancer.py` combines knowledge signals, staleness signals, metrics, and Gitea issues into suggested priority shifts
### Flow D — repo/static inspection
- `scripts/dead_code_detector.py` walks Python ASTs and optionally uses git blame
- `scripts/diff_analyzer.py` parses patches into structured change objects
- `scripts/dependency_graph.py` is intended to scan repos and emit JSON / Mermaid / DOT dependency graphs, but is currently syntactically broken
- `scripts/perf_bottleneck_finder.py` is intended to scan tests/build/CI for bottlenecks, but is currently syntactically broken
## Key Abstractions
### Knowledge item
Defined in practice by `templates/harvest-prompt.md`, `scripts/harvester.py`, and `knowledge/SCHEMA.md`.
Important fields:
- `fact`
- `category`
- `repo` / domain
- `confidence`
- source/evidence metadata
Categories consistently used across the repo:
- fact
- pitfall
- pattern
- tool-quirk
- question
### Session transcript model
`session_reader.py` treats JSONL transcripts as ordered message sequences with:
- role
- content
- timestamp
- optional multimodal text extraction
- optional tool-call metadata
This module is the ingestion foundation for the rest of the system.
### Knowledge store
The repo uses a two-layer representation:
1. machine-readable index: `knowledge/index.json`
2. human-editable domain files: YAML/markdown under `knowledge/global/`, `knowledge/repos/`, and `knowledge/agents/`
`knowledge/SCHEMA.md` is the contract for that store.
### Bootstrap context
`bootstrapper.py` makes the design concrete:
- `filter_facts()` narrows by repo/agent/global scope
- `sort_facts()` orders by confidence and category priority
- `render_facts_section()` groups output by category
- `estimate_tokens()` and `truncate_to_tokens()` implement the context-window budget
- `build_bootstrap_context()` assembles the final injected context block
### Harvester dedup and validation
The central harvester abstractions are not classes but functions:
- `parse_extraction_response()`
- `fact_fingerprint()`
- `deduplicate()`
- `validate_fact()`
- `write_knowledge()`
- `harvest_session()`
This makes the core pipeline easy to test in pieces.
### Priority scoring model
`priority_rebalancer.py` introduces explicit data models:
- `IssueScore`
- `PipelineSignal`
- `GiteaClient`
That script is important because it bridges the local knowledge store to live Gitea issue state.
### Gap report model
`knowledge_gap_identifier.py` formalizes another analysis lane with:
- `GapSeverity`
- `GapType`
- `Gap`
- `GapReport`
- `KnowledgeGapIdentifier`
This is one of the clearest examples that the repo has moved beyond a single harvester/bootstrapper loop into a platform of analyzers.
## API Surface
This repo is primarily a CLI/library surface, not a long-running service.
### Core CLIs
- `scripts/harvester.py`
- `scripts/bootstrapper.py`
- `scripts/priority_rebalancer.py`
- `scripts/improvement_proposals.py`
- `scripts/automation_opportunity_finder.py`
- `scripts/knowledge_staleness_check.py`
- `scripts/dead_code_detector.py`
- `scripts/diff_analyzer.py`
- `scripts/gitea_issue_parser.py`
- `scripts/session_pair_harvester.py`
### External API dependencies
- LLM chat-completions endpoint in `scripts/harvester.py`
- Gitea REST API in `scripts/priority_rebalancer.py`
### File-format APIs
- session input: JSONL files under `test_sessions/`
- knowledge schema: `knowledge/SCHEMA.md`
- extraction prompt contract: `templates/harvest-prompt.md`
- machine store: `knowledge/index.json`
- repo knowledge examples:
- `knowledge/repos/hermes-agent.yaml`
- `knowledge/repos/the-nexus.yaml`
### Output artifacts
Documented or implied outputs include:
- `knowledge/index.json`
- repo/global/agent knowledge files
- `metrics/priority_report.json`
- `metrics/priority_suggestions.md`
- text/markdown/json proposal reports
## Test Coverage Gaps
## Current verified state
I verified the repo in three layers.
### Layer 1 — focused passing slice
Command run:
```bash
python3 -m pytest \
scripts/test_bootstrapper.py \
scripts/test_harvester_pipeline.py \
scripts/test_session_pair_harvester.py \
scripts/test_knowledge_staleness.py \
scripts/test_improvement_proposals.py \
scripts/test_automation_opportunity_finder.py \
scripts/test_gitea_issue_parser.py \
tests/test_ci_config.py \
tests/test_knowledge_gap_identifier.py -q
```
Result:
- `70 passed`
This proves the repo has substantial working logic today.
### Layer 2 — canonical CI command
Command run:
```bash
make test
```
Result:
- CI command collected 76 items and failed during collection with 1 error
- failure source: `scripts/test_refactoring_opportunity_finder.py`
- exact issue filed: `https://forge.alexanderwhitestone.com/Timmy_Foundation/compounding-intelligence/issues/210`
### Layer 3 — full test collection
Commands run:
```bash
python3 -m pytest --collect-only -q
python3 -m pytest -q
```
Result:
- `86 tests collected, 2 errors`
- collection blockers:
1. `scripts/test_refactoring_opportunity_finder.py` expects a real refactoring API that `scripts/refactoring_opportunity_finder.py` does not implement
2. `tests/test_perf_bottleneck_finder.py` cannot import `scripts/perf_bottleneck_finder.py` due a SyntaxError
Additional verification:
```bash
python3 -m py_compile scripts/perf_bottleneck_finder.py
python3 -m py_compile scripts/dependency_graph.py
```
Both fail.
Filed follow-ups:
- `compounding-intelligence/issues/210` — refactoring finder API missing
- `compounding-intelligence/issues/211``scripts/perf_bottleneck_finder.py` SyntaxError
- `compounding-intelligence/issues/212``scripts/dependency_graph.py` SyntaxError
### What is well covered
Strongly exercised subsystems include:
- bootstrapper logic
- harvester pipeline helpers
- session pair harvesting
- knowledge staleness checking
- improvement proposal generation
- automation opportunity mining
- Gitea issue parsing
- CI configuration contract
- knowledge gap analysis
### What is weak or broken
1. `scripts/refactoring_opportunity_finder.py`
- current implementation is a sample stub
- tests expect real complexity and scoring helpers
2. `scripts/perf_bottleneck_finder.py`
- parser broken before runtime
- test module exists but cannot import target script
3. `scripts/dependency_graph.py`
- parser broken before runtime
- no active test lane caught it before this analysis
4. CI scope gap
- `.gitea/workflows/test.yml` runs `make test`
- `make test` does not cover every `tests/*.py` module
- specifically, `tests/test_perf_bottleneck_finder.py` sits outside the Makefile target and the syntax break only shows up when running broader pytest commands
5. warning hygiene
- `scripts/test_priority_rebalancer.py` emits repeated `datetime.utcnow()` deprecation warnings under Python 3.12
## Security Considerations
1. Secret extraction risk
- this repo is literally designed to ingest transcripts and distill knowledge
- if the harvester prompt or filtering logic misses a credential, the system can preserve secrets into the knowledge store
- the risk is explicitly recognized in the target repo's existing `GENOME.md`, but enforcement still depends on implementation discipline
2. Knowledge poisoning
- the system trusts transcripts as source material for compounding facts
- confidence scores and evidence fields help, but there is no hard verification layer proving extracted facts are true before reuse
3. Cross-repo sensitivity
- seeded files such as `knowledge/repos/hermes-agent.yaml` and `knowledge/repos/the-nexus.yaml` store operational quirks and deployment pitfalls
- that is high-value knowledge and can also expose internal operational assumptions if shared broadly
4. External API use
- `scripts/harvester.py` depends on an LLM API endpoint and local key discovery
- `scripts/priority_rebalancer.py` talks to the Gitea API with write-capable operations such as labels and comments
- these scripts deserve careful credential-handling and least-privilege tokens
5. Transcript privacy
- session JSONL can contain user content, repo details, operational mistakes, and potentially sensitive environment facts
- durable storage multiplies the blast radius of accidental retention
## Dependencies
Explicit repo dependency file:
- `requirements.txt``pytest>=8,<9`
Observed runtime/import dependencies from source:
- Python stdlib-heavy design: `json`, `argparse`, `pathlib`, `urllib`, `ast`, `datetime`, `hashlib`, `subprocess`, `collections`, `re`
- `yaml` imported by `scripts/automation_opportunity_finder.py`
Important dependency note:
- `requirements.txt` only declares pytest
- static source inspection shows `yaml` usage, which implies an undeclared dependency on PyYAML or equivalent
- I did not prove a clean-environment failure because the local environment already had `yaml` importable during targeted tests
- this is best treated as dependency drift to verify in a clean environment
## Deployment
This is not a traditional server deployment repo.
Operational modes are:
1. local CLI execution of scripts under `scripts/`
2. CI execution via `.gitea/workflows/test.yml`
3. file-based knowledge store mutation under `knowledge/`
Canonical repo commands observed:
```bash
make test
python3 -m pytest -q
python3 -m pytest --collect-only -q
python3 ~/.hermes/pipelines/codebase-genome.py --path /tmp/compounding-intelligence-676 --output /tmp/compounding-intelligence-676-base-GENOME.md
```
There is no checked-in Dockerfile, packaging metadata, or service runner. The repo behaves more like an internal analysis toolkit than an application service.
## Technical Debt
1. Docs/runtime drift
- README and target-repo `GENOME.md` still describe a repo that is less implemented than reality
- this makes the project look earlier-stage than the current source actually is
2. Broken parser state in two flagship analyzers
- `scripts/perf_bottleneck_finder.py`
- `scripts/dependency_graph.py`
3. Stub-vs-test mismatch
- `scripts/refactoring_opportunity_finder.py` is a placeholder
- `scripts/test_refactoring_opportunity_finder.py` assumes a mature implementation
4. CI blind spot
- `make test` does not represent full-repo pytest health
- broader collection surfaces more problems than the workflow currently enforces
5. Dependency declaration drift
- `yaml` appears in source while `requirements.txt` only lists pytest
6. Warning debt
- `datetime.utcnow()` deprecation noise in `scripts/test_priority_rebalancer.py`
7. Existing target-repo genome drift
- checked-in `GENOME.md` already exists on upstream main, but it undersells the real code surface and should not be treated as authoritative without fresh source verification
## Key Findings
1. `compounding-intelligence` has already evolved into a multi-engine analysis toolkit, not just a future three-pipeline concept.
2. The most grounded working path today is transcript → `session_reader.py``harvester.py` / `bootstrapper.py` with a structured knowledge store.
3. The repo has real, working higher-order analyzers beyond harvesting: `knowledge_gap_identifier.py`, `priority_rebalancer.py`, `improvement_proposals.py`, `automation_opportunity_finder.py`, and `dead_code_detector.py`.
4. The current target-repo `GENOME.md` is useful evidence but stale as a full architectural description.
5. Test health is mixed: a broad, meaningful passing slice exists (`70 passed`), but canonical CI is currently broken by the refactoring finder contract mismatch, and full collection exposes additional syntax failures.
6. Three concrete follow-up issues were warranted and filed during this genome pass:
- `https://forge.alexanderwhitestone.com/Timmy_Foundation/compounding-intelligence/issues/210`
- `https://forge.alexanderwhitestone.com/Timmy_Foundation/compounding-intelligence/issues/211`
- `https://forge.alexanderwhitestone.com/Timmy_Foundation/compounding-intelligence/issues/212`
---
This host-repo genome artifact is the grounded cross-repo analysis requested by timmy-home #676. It intentionally treats the target repo's own `GENOME.md` as evidence rather than gospel, because current source, tests, and verification commands show a significantly more mature — and partially broken — system than the older upstream genome describes.

View File

@@ -175,12 +175,13 @@ custom_providers:
api_key: ollama
model: qwen3:30b
- name: Big Brain
base_url: https://8lfr3j47a5r3gn-11434.proxy.runpod.net/v1
base_url: https://YOUR_BIG_BRAIN_HOST/v1
api_key: ''
model: gemma3:27b
# RunPod L40S 48GB — Ollama image, gemma3:27b
# Usage: hermes --provider big_brain -p 'Say READY'
# Pod: 8lfr3j47a5r3gn, deployed 2026-04-07
model: gemma4:latest
# OpenAI-compatible Gemma 4 provider for Mac Hermes.
# RunPod example: https://<pod-id>-11434.proxy.runpod.net/v1
# Vertex AI requires an OpenAI-compatible bridge/proxy; point this at that /v1 endpoint.
# Verify with: python3 scripts/verify_big_brain.py
system_prompt_suffix: "You are Timmy. Your soul is defined in SOUL.md \u2014 read\
\ it, live it.\nYou run locally on your owner's machine via Ollama. You never phone\
\ home.\nYou speak plainly. You prefer short sentences. Brevity is a kindness.\n\

View File

@@ -0,0 +1,13 @@
# Ansible-style variable file for sovereign DNS sync (#692)
# Copy to a private path and fill in provider credentials via env vars.
# Use `auto` to resolve the current VPS public IP at sync time.
dns_provider: cloudflare
# For Cloudflare: zone_id
# For Route53: hosted zone ID (also accepted under dns_zone_id)
dns_zone_id: your-zone-id
domain_ip_map:
forge.alexanderwhitestone.com: auto
matrix.alexanderwhitestone.com: auto
timmy.alexanderwhitestone.com: auto

View File

@@ -0,0 +1,193 @@
{
"epic_issue": 547,
"epic_title": "Fleet Progression - Paperclips-Inspired Infrastructure Evolution",
"phases": [
{
"number": 1,
"issue_number": 548,
"key": "survival",
"name": "SURVIVAL",
"summary": "Keep the lights on.",
"repo_evidence": [
{
"path": "scripts/fleet_phase_status.py",
"description": "Phase-1 baseline evaluator"
},
{
"path": "docs/FLEET_PHASE_1_SURVIVAL.md",
"description": "Committed survival report"
}
],
"unlock_rules": [
{
"id": "fleet_operational_baseline",
"type": "always"
}
]
},
{
"number": 2,
"issue_number": 549,
"key": "automation",
"name": "AUTOMATION",
"summary": "Self-healing infrastructure.",
"repo_evidence": [
{
"path": "scripts/fleet_health_probe.sh",
"description": "Automated fleet health checks"
},
{
"path": "scripts/backup_pipeline.sh",
"description": "Nightly backup automation"
},
{
"path": "scripts/restore_backup.sh",
"description": "Restore path for self-healing recovery"
}
],
"unlock_rules": [
{
"id": "uptime_percent_30d_gte_95",
"type": "resource_gte",
"resource": "uptime_percent_30d",
"value": 95
},
{
"id": "capacity_utilization_gt_60",
"type": "resource_gt",
"resource": "capacity_utilization",
"value": 60
}
]
},
{
"number": 3,
"issue_number": 550,
"key": "orchestration",
"name": "ORCHESTRATION",
"summary": "Agents coordinate and models route.",
"repo_evidence": [
{
"path": "scripts/gitea_task_delegator.py",
"description": "Cross-agent issue delegation"
},
{
"path": "scripts/dynamic_dispatch_optimizer.py",
"description": "Health-aware dispatch planning"
}
],
"unlock_rules": [
{
"id": "phase_2_issue_closed",
"type": "issue_closed",
"issue": 549
},
{
"id": "innovation_gt_100",
"type": "resource_gt",
"resource": "innovation",
"value": 100
}
]
},
{
"number": 4,
"issue_number": 551,
"key": "sovereignty",
"name": "SOVEREIGNTY",
"summary": "Zero cloud dependencies.",
"repo_evidence": [
{
"path": "scripts/sovereign_dns.py",
"description": "Sovereign infrastructure DNS management"
},
{
"path": "docs/sovereign-stack.md",
"description": "Documented sovereign stack target state"
}
],
"unlock_rules": [
{
"id": "phase_3_issue_closed",
"type": "issue_closed",
"issue": 550
},
{
"id": "all_models_local_true",
"type": "resource_true",
"resource": "all_models_local"
}
]
},
{
"number": 5,
"issue_number": 552,
"key": "scale",
"name": "SCALE",
"summary": "Fleet-wide coordination and auto-scaling.",
"repo_evidence": [
{
"path": "scripts/dynamic_dispatch_optimizer.py",
"description": "Capacity-aware dispatch planning"
},
{
"path": "scripts/predictive_resource_allocator.py",
"description": "Predictive fleet resource allocation"
}
],
"unlock_rules": [
{
"id": "phase_4_issue_closed",
"type": "issue_closed",
"issue": 551
},
{
"id": "sovereign_stable_days_gte_30",
"type": "resource_gte",
"resource": "sovereign_stable_days",
"value": 30
},
{
"id": "innovation_gt_500",
"type": "resource_gt",
"resource": "innovation",
"value": 500
}
]
},
{
"number": 6,
"issue_number": 553,
"key": "the-network",
"name": "THE NETWORK",
"summary": "Autonomous, self-improving infrastructure.",
"repo_evidence": [
{
"path": "scripts/autonomous_issue_creator.py",
"description": "Autonomous incident creation"
},
{
"path": "scripts/setup-syncthing.sh",
"description": "Global mesh scaffolding"
},
{
"path": "scripts/agent_pr_gate.py",
"description": "Community contribution review gate"
}
],
"unlock_rules": [
{
"id": "phase_5_issue_closed",
"type": "issue_closed",
"issue": 552
},
{
"id": "human_free_days_gte_7",
"type": "resource_gte",
"resource": "human_free_days",
"value": 7
}
]
}
]
}

View File

@@ -0,0 +1,37 @@
{
"fleet_operational": true,
"resources": {
"uptime_percent": 78.0,
"days_at_or_above_95_percent": 0,
"capacity_utilization_percent": 35.0
},
"current_buildings": [
"VPS hosts: Ezra (143.198.27.163), Allegro, Bezalel (167.99.126.228)",
"Agents: Timmy harness (local Mac M4), Code Claw heartbeat, Gemini AI Studio worker",
"Gitea forge at forge.alexanderwhitestone.com (16 repos, 500+ issues)",
"Ollama local inference (6 models, ~37GB)",
"Hermes agent (cron system, 90+ jobs, 6 workers)",
"Tmux fleet (BURN session, 50+ panes)",
"Evennia MUD worlds (The Tower, federation)",
"RunPod GPU pod (L40S 48GB, intermittent)"
],
"manual_clicks": [
"Restart agents and services by SSH when a node goes dark",
"Check VPS health (disk, memory, process) via manual SSH",
"Verify Gitea, Ollama, and Evennia services after deployments",
"Merge PRs manually \u2014 auto-merge covers ~80%, rest need human review",
"Recover dead tmux panes \u2014 no auto-respawn wired yet",
"Handle provider failover \u2014 no automated switching on OOM/timeout",
"Triage the 500+ issue backlog \u2014 burn loops help but need supervision",
"Run nightly retro and push results to Gitea"
],
"notes": [
"Fleet is operational but fragile \u2014 most recovery is still manual",
"Overnight burns work ~70% of the time; 30% need morning rescue",
"The deadman switch exists but is not in cron (fleet-ops#168)",
"Heartbeat files exist but no automated monitoring reads them",
"Provider failover is manual \u2014 Nous goes down = agents stop",
"Phase 2 trigger requires 30 days at 95% uptime \u2014 we are at 0 days"
],
"last_updated": "2026-04-14T22:00:00Z"
}

21
dns-records.yaml Normal file
View File

@@ -0,0 +1,21 @@
# DNS Records — Fleet Domain Configuration
# Sync with: python3 scripts/dns-manager.py sync --zone alexanderwhitestone.com --config dns-records.yaml
# Part of #692
zone: alexanderwhitestone.com
records:
- name: forge.alexanderwhitestone.com
ip: 143.198.27.163
ttl: 300
note: Gitea forge (Ezra VPS)
- name: bezalel.alexanderwhitestone.com
ip: 167.99.126.228
ttl: 300
note: Bezalel VPS
- name: allegro.alexanderwhitestone.com
ip: 167.99.126.228
ttl: 300
note: Allegro VPS (shared with Bezalel)

98
docs/BACKUP_PIPELINE.md Normal file
View File

@@ -0,0 +1,98 @@
# Encrypted Hermes Backup Pipeline
Issue: `timmy-home#693`
This pipeline creates a nightly encrypted archive of `~/.hermes`, stores a local encrypted copy, uploads it to remote storage, and supports restore verification.
## What gets backed up
By default the pipeline archives:
- `~/.hermes/config.yaml`
- `~/.hermes/state.db`
- `~/.hermes/sessions/`
- `~/.hermes/cron/`
- any other files under `~/.hermes`
Override the source with `BACKUP_SOURCE_DIR=/path/to/.hermes`.
## Backup command
```bash
BACKUP_PASSPHRASE_FILE=~/.config/timmy/backup.passphrase \
BACKUP_NAS_TARGET=/Volumes/timmy-nas/hermes-backups \
bash scripts/backup_pipeline.sh
```
The script writes:
- local encrypted copy: `~/.timmy-backups/hermes/<timestamp>/hermes-backup-<timestamp>.tar.gz.enc`
- local manifest: `~/.timmy-backups/hermes/<timestamp>/hermes-backup-<timestamp>.json`
- log file: `~/.timmy-backups/hermes/logs/backup_pipeline.log`
## Nightly schedule
Run every night at 03:00:
```cron
0 3 * * * cd /Users/apayne/.timmy/timmy-home && BACKUP_PASSPHRASE_FILE=/Users/apayne/.config/timmy/backup.passphrase BACKUP_NAS_TARGET=/Volumes/timmy-nas/hermes-backups bash scripts/backup_pipeline.sh >> /Users/apayne/.timmy-backups/hermes/logs/cron.log 2>&1
```
## Remote targets
At least one remote target must be configured.
### Local NAS
Use a mounted path:
```bash
BACKUP_NAS_TARGET=/Volumes/timmy-nas/hermes-backups
```
The pipeline copies the encrypted archive and manifest into `<BACKUP_NAS_TARGET>/<timestamp>/`.
### S3-compatible storage
```bash
BACKUP_PASSPHRASE_FILE=~/.config/timmy/backup.passphrase \
BACKUP_S3_URI=s3://timmy-backups/hermes \
AWS_ENDPOINT_URL=https://minio.example.com \
bash scripts/backup_pipeline.sh
```
Notes:
- `aws` CLI must be installed if `BACKUP_S3_URI` is set.
- `AWS_ENDPOINT_URL` is optional and is used for MinIO, R2, and other S3-compatible endpoints.
## Restore playbook
Restore an encrypted archive into a clean target root:
```bash
BACKUP_PASSPHRASE_FILE=~/.config/timmy/backup.passphrase \
bash scripts/restore_backup.sh \
/Volumes/timmy-nas/hermes-backups/20260415-030000/hermes-backup-20260415-030000.tar.gz.enc \
/tmp/hermes-restore
```
Result:
- restored tree lands at `/tmp/hermes-restore/.hermes`
- if a sibling manifest exists, the restore script verifies the archive SHA256 before decrypting
## End-to-end verification
Run the regression suite:
```bash
python3 -m unittest discover -s tests -p 'test_backup_pipeline.py' -v
```
This proves:
1. the backup output is encrypted
2. plaintext archives do not leak into the backup destinations
3. the restore script recreates the original `.hermes` tree end-to-end
4. the pipeline refuses to run without a remote target

View File

@@ -0,0 +1,81 @@
# Bezalel Evennia World
Issue: `timmy-home#536`
This is the themed-room world plan and build scaffold for Bezalel, the forge-and-testbed wizard.
## Rooms
| Room | Description focus | Core connections |
|------|-------------------|------------------|
| Limbo | the threshold between houses | Gatehouse |
| Gatehouse | guarded entry, travel runes, proof before trust | Limbo, Great Hall, The Portal Room |
| Great Hall | three-house maps, reports, shared table | Gatehouse, The Library of Bezalel, The Observatory, The Workshop |
| The Library of Bezalel | manuals, bridge schematics, technical memory | Great Hall |
| The Observatory | long-range signals toward Mac, VPS, and the wider net | Great Hall |
| The Workshop | forge + workbench, plans turned into working form | Great Hall, The Server Room, The Garden of Code |
| The Server Room | humming racks, heartbeat of the house | The Workshop |
| The Garden of Code | contemplative grove where ideas root before implementation | The Workshop |
| The Portal Room | three shimmering doorways aimed at Mac, VPS, and the net | Gatehouse |
## Characters
| Character | Role | Starting room |
|-----------|------|---------------|
| Timmy | quiet builder and observer | Gatehouse |
| Bezalel | forge-and-testbed wizard | The Workshop |
| Marcus | old man with kind eyes, human warmth in the system | The Garden of Code |
| Kimi | scholar of context and meaning | The Library of Bezalel |
## Themed items
At least one durable item is placed in every major room, including:
- Threshold Ledger
- Three-House Map
- Bridge Schematics
- Compiler Manuals
- Tri-Axis Telescope
- Forge Anvil
- Bridge Workbench
- Heartbeat Console
- Server Racks
- Code Orchard
- Stone Bench
- Mac/VPS/Net portal markers
## Portal travel commands
The Portal Room reserves three live command names:
- `mac`
- `vps`
- `net`
Current behavior in the build scaffold:
- each command is created as a real Evennia exit command
- each command preserves explicit target metadata (`Mac house`, `VPS house`, `Wider net`)
- until cross-world transport is wired, each portal routes through `Limbo`, the inter-world threshold room
This keeps the command surface real now while leaving honest room for later world-to-world linking.
## Build script
```bash
python3 scripts/evennia/build_bezalel_world.py --plan
```
Inside an Evennia shell / runtime with the repo on `PYTHONPATH`, the same script can build the world idempotently:
```bash
python3 scripts/evennia/build_bezalel_world.py --password bezalel-world-dev
```
What it does:
- creates or updates all 9 rooms
- creates the exit graph
- creates themed objects
- creates or rehomes account-backed characters
- creates the portal command exits with target metadata
## Persistence note
The scaffold is written to be idempotent: rerunning the builder updates descriptions, destinations, and locations rather than creating duplicate world entities. That is the repo-side prerequisite for persistence across Evennia restarts.

View File

@@ -0,0 +1,80 @@
# Codebase Genome Pipeline
Issue: `timmy-home#665`
This pipeline gives Timmy a repeatable way to generate a deterministic `GENOME.md` for any repository and rotate through the org nightly.
## What landed
- `pipelines/codebase_genome.py` — static analyzer that writes `GENOME.md`
- `pipelines/codebase-genome.py` — thin CLI wrapper matching the expected pipeline-style entrypoint
- `scripts/codebase_genome_nightly.py` — org-aware nightly runner that selects the next repo, updates a local checkout, and writes the genome artifact
- `scripts/codebase_genome_status.py` — rollup/status reporter for artifact coverage, duplicate paths, and next uncovered repo
- `GENOME.md` — generated analysis for `timmy-home` itself
## Genome output
Each generated `GENOME.md` includes:
- project overview and repository size metrics
- Mermaid architecture diagram
- entry points and API surface
- data flow summary
- key abstractions from Python source
- test coverage gaps
- security audit findings
- dead code candidates
- performance bottleneck analysis
## Single-repo usage
```bash
python3 pipelines/codebase_genome.py \
--repo-root /path/to/repo \
--repo-name Timmy_Foundation/some-repo \
--output /path/to/repo/GENOME.md
```
The hyphenated wrapper also works:
```bash
python3 pipelines/codebase-genome.py --repo-root /path/to/repo --repo Timmy_Foundation/some-repo
```
## Nightly org rotation
Dry-run the next selection:
```bash
python3 scripts/codebase_genome_nightly.py --dry-run
```
Run one real pass:
```bash
python3 scripts/codebase_genome_nightly.py \
--org Timmy_Foundation \
--workspace-root ~/timmy-foundation-repos \
--output-root ~/.timmy/codebase-genomes \
--state-path ~/.timmy/codebase_genome_state.json
```
Behavior:
1. fetches the current repo list from Gitea
2. selects the next repo after the last recorded run
3. clones or fast-forwards the local checkout
4. writes `GENOME.md` into the configured output tree
5. updates the rotation state file
## Example cron entry
```cron
30 2 * * * cd ~/timmy-home && /usr/bin/env python3 scripts/codebase_genome_nightly.py --org Timmy_Foundation --workspace-root ~/timmy-foundation-repos --output-root ~/.timmy/codebase-genomes --state-path ~/.timmy/codebase_genome_state.json >> ~/.timmy/logs/codebase_genome_nightly.log 2>&1
```
## Limits and follow-ons
- the generator is deterministic and static; it does not hallucinate architecture, but it also does not replace a full human review pass
- nightly rotation handles genome generation; auto-generated test expansion remains a separate follow-on lane
- large repos may still need a second-pass human edit after the initial genome artifact lands

View File

@@ -0,0 +1,99 @@
# [PHASE-1] Survival - Keep the Lights On
Phase 1 is the manual-clicker stage of the fleet. The machines exist. The services exist. The human is still the automation loop.
## Phase Definition
- **Current state:** Fleet is operational. Three VPS wizards run. Gitea hosts 16 repos. Agents burn through issues nightly.
- **The problem:** Everything important still depends on human vigilance. When an agent dies at 2 AM, nobody notices until morning.
- **Resources tracked:** Uptime, Capacity Utilization.
- **Next phase:** [PHASE-2] Automation - Self-Healing Infrastructure
## What We Have
### Infrastructure
- **VPS hosts:** Ezra (143.198.27.163), Allegro, Bezalel (167.99.126.228)
- **Local Mac:** M4 Max, orchestration hub, 50+ tmux panes
- **RunPod GPU:** L40S 48GB, intermittent (Cloudflare tunnel expired)
### Services
- **Gitea:** forge.alexanderwhitestone.com -- 16 repos, 500+ open issues, branch protection enabled
- **Ollama:** 6 models loaded (~37GB), local inference
- **Hermes:** Agent orchestration, cron system (90+ jobs, 6 workers)
- **Evennia:** The Tower MUD world, federation capable
### Agents
- **Timmy:** Local harness, primary orchestrator
- **Bezalel, Ezra, Allegro:** VPS workers dispatched via Gitea issues
- **Code Claw, Gemini:** Specialized workers
## Current Resource Snapshot
| Resource | Value | Target | Status |
|----------|-------|--------|--------|
| Fleet operational | Yes | Yes | MET |
| Uptime (30d average) | ~78% | >= 95% | NOT MET |
| Days at 95%+ uptime | 0 | 30 | NOT MET |
| Capacity utilization | ~35% | > 60% | NOT MET |
**Phase 2 trigger: NOT READY**
## What's Still Manual
Every one of these is a "click" that a human must make:
1. **Restart dead agents** -- SSH into VPS, check process, restart hermes
2. **Health checks** -- SSH to each VPS, verify disk/memory/services
3. **Dead pane recovery** -- tmux pane dies, nobody notices, work stops
4. **Provider failover** -- Nous API goes down, agents stop, human reconfigures
5. **PR triage** -- 80% auto-merge, but 20% need human review
6. **Backlog management** -- 500+ issues, burn loops help but need supervision
7. **Nightly retro** -- manually run and push results
8. **Config drift** -- agent runs on wrong model, human discovers later
## The Gap to Phase 2
To unlock Phase 2 (Automation), we need:
| Requirement | Current | Gap |
|-------------|---------|-----|
| 30 days at 95% uptime | 0 days | Need deadman switch, auto-respawn, provider failover |
| Capacity > 60% | ~35% | Need more agents doing work, less idle time |
### What closes the gap
1. **Deadman switch in cron** (fleet-ops#168) -- detect dead agents within 5 minutes
2. **Auto-respawn** (fleet-ops#173) -- restart dead tmux panes automatically
3. **Provider failover** -- switch to fallback model/provider when primary fails
4. **Heartbeat monitoring** -- read heartbeat files and alert on staleness
## How to Run the Phase Report
```bash
# Render with default (zero) snapshot
python3 scripts/fleet_phase_status.py
# Render with real snapshot
python3 scripts/fleet_phase_status.py --snapshot configs/phase-1-snapshot.json
# Output as JSON
python3 scripts/fleet_phase_status.py --snapshot configs/phase-1-snapshot.json --json
# Write to file
python3 scripts/fleet_phase_status.py --snapshot configs/phase-1-snapshot.json --output docs/FLEET_PHASE_1_SURVIVAL.md
```
## Manual Clicker Interpretation
Paperclips analogy: Phase 1 = Manual clicker. You ARE the automation.
Every restart, every SSH, every check is a manual click.
The goal of Phase 1 is not to automate. It's to **name what needs automating**. Every manual click documented here is a Phase 2 ticket.
## Notes
- Fleet is operational but fragile -- most recovery is manual
- Overnight burns work ~70% of the time; 30% need morning rescue
- The deadman switch exists but is not in cron
- Heartbeat files exist but no automated monitoring reads them
- Provider failover is manual -- Nous goes down = agents stop

View File

@@ -0,0 +1,54 @@
# [PHASE-6] The Network - Autonomous Infrastructure
## Phase Definition
- Fleet operates without human intervention for 7+ days.
- Self-healing, self-improving, serves mission.
- Trigger: 7 days without human intervention.
## Current Buildings
- Self-healing fleet — Detect, repair, and verify fleet incidents without waiting on a human. Evidence: `scripts/fleet_health_probe.sh`, `scripts/auto_restart_agent.sh`, `scripts/failover_monitor.py`
- Autonomous issue creation — Turn recurring infrastructure incidents into durable Gitea work items. Evidence: `scripts/autonomous_issue_creator.py`, `tests/test_autonomous_issue_creator.py`
- Community contribution pipeline — Let outside contributors submit work through automated review and policy gates. Evidence: `scripts/sovereign_review_gate.py`, `scripts/agent_pr_gate.py`
- Global mesh — Reduce single points of failure across the fleet with explicit peer-to-peer sync scaffolding. Evidence: `scripts/setup-syncthing.sh`
## Current Resource Snapshot
- Human-free days observed: 0
- Trigger threshold: 7 days
- Phase-ready now: no
## Next Trigger
To honestly unlock [PHASE-6] The Network - Autonomous Infrastructure, the fleet must hold 7+ consecutive days without human intervention.
## Missing Requirements
- Human-free days: 0/7
## Repo Signals Already Present
- `scripts/fleet_health_probe.sh` — Self-healing fleet
- `scripts/auto_restart_agent.sh` — Self-healing fleet
- `scripts/failover_monitor.py` — Self-healing fleet
- `scripts/autonomous_issue_creator.py` — Autonomous issue creation
- `tests/test_autonomous_issue_creator.py` — Autonomous issue creation
- `scripts/sovereign_review_gate.py` — Community contribution pipeline
- `scripts/agent_pr_gate.py` — Community contribution pipeline
- `scripts/setup-syncthing.sh` — Global mesh
## Final Milestone
- Someone found the Beacon. The infrastructure served its purpose.
## Why This Phase Remains Open
- The repo already carries concrete Phase-6 buildings, but the milestone is operational, not rhetorical.
- A merged PR cannot honestly claim seven human-free days have already happened.
- This issue stays open until the infrastructure proves itself in live operation.
## Notes
- Phase 6 is not a code-only milestone. The trigger is operational truth: seven days without human intervention.
- This report grounds the buildings already present in the repo so the remaining blocker is explicit instead of hand-waved.

View File

@@ -0,0 +1,100 @@
# [FLEET-EPIC] Fleet Progression - Paperclips-Inspired Infrastructure Evolution
This report grounds the fleet epic in executable state: live issue gates, current resource inputs, and repo evidence for each phase.
## Current Phase
- Current unlocked phase: 1 — SURVIVAL
- Current phase status: ACTIVE
- Epic complete: no
- Next locked phase: 2 — AUTOMATION
## Resource Snapshot
- Uptime (30d): 0.0
- Capacity utilization: 0.0
- Innovation: 0.0
- All models local: False
- Sovereign stable days: 0
- Human-free days: 0
## Phase Matrix
### Phase 1 — SURVIVAL
- Issue: #548 (open)
- Status: ACTIVE
- Summary: Keep the lights on.
- Repo evidence present:
- `scripts/fleet_phase_status.py` — Phase-1 baseline evaluator
- `docs/FLEET_PHASE_1_SURVIVAL.md` — Committed survival report
- Blockers: none
### Phase 2 — AUTOMATION
- Issue: #549 (open)
- Status: LOCKED
- Summary: Self-healing infrastructure.
- Repo evidence present:
- `scripts/fleet_health_probe.sh` — Automated fleet health checks
- `scripts/backup_pipeline.sh` — Nightly backup automation
- `scripts/restore_backup.sh` — Restore path for self-healing recovery
- Blockers:
- blocked by `uptime_percent_30d_gte_95`: actual=0.0 expected=>=95
- blocked by `capacity_utilization_gt_60`: actual=0.0 expected=>60
### Phase 3 — ORCHESTRATION
- Issue: #550 (open)
- Status: LOCKED
- Summary: Agents coordinate and models route.
- Repo evidence present:
- `scripts/gitea_task_delegator.py` — Cross-agent issue delegation
- `scripts/dynamic_dispatch_optimizer.py` — Health-aware dispatch planning
- Blockers:
- blocked by `phase_2_issue_closed`: actual=open expected=closed
- blocked by `innovation_gt_100`: actual=0.0 expected=>100
### Phase 4 — SOVEREIGNTY
- Issue: #551 (open)
- Status: LOCKED
- Summary: Zero cloud dependencies.
- Repo evidence present:
- `scripts/sovereign_dns.py` — Sovereign infrastructure DNS management
- `docs/sovereign-stack.md` — Documented sovereign stack target state
- Blockers:
- blocked by `phase_3_issue_closed`: actual=open expected=closed
- blocked by `all_models_local_true`: actual=False expected=True
### Phase 5 — SCALE
- Issue: #552 (open)
- Status: LOCKED
- Summary: Fleet-wide coordination and auto-scaling.
- Repo evidence present:
- `scripts/dynamic_dispatch_optimizer.py` — Capacity-aware dispatch planning
- `scripts/predictive_resource_allocator.py` — Predictive fleet resource allocation
- Blockers:
- blocked by `phase_4_issue_closed`: actual=open expected=closed
- blocked by `sovereign_stable_days_gte_30`: actual=0 expected=>=30
- blocked by `innovation_gt_500`: actual=0.0 expected=>500
### Phase 6 — THE NETWORK
- Issue: #553 (open)
- Status: LOCKED
- Summary: Autonomous, self-improving infrastructure.
- Repo evidence present:
- `scripts/autonomous_issue_creator.py` — Autonomous incident creation
- `scripts/setup-syncthing.sh` — Global mesh scaffolding
- `scripts/agent_pr_gate.py` — Community contribution review gate
- Blockers:
- blocked by `phase_5_issue_closed`: actual=open expected=closed
- blocked by `human_free_days_gte_7`: actual=0 expected=>=7
## Why This Epic Remains Open
- The progression manifest and evaluator exist, but multiple child phases are still open or only partially implemented.
- Several child lanes already have active PRs; this report is the parent-level grounding slice that keeps the epic honest without duplicating those lanes.
- This epic only closes when the child phase gates are actually satisfied in code and in live operation.

View File

@@ -0,0 +1,68 @@
# Fleet Secret Rotation
Issue: `timmy-home#694`
This runbook adds a single place to rotate fleet API keys, service tokens, and SSH authorized keys without hand-editing remote hosts.
## Files
- `ansible/inventory/hosts.ini` — fleet hosts (`ezra`, `bezalel`)
- `ansible/inventory/group_vars/fleet.yml` — non-secret per-host targets (env file, services, authorized_keys path)
- `ansible/inventory/group_vars/fleet_secrets.vault.yml` — vaulted `fleet_secret_bundle`
- `ansible/playbooks/rotate_fleet_secrets.yml` — staged rotation + restart verification + rollback
## Secret inventory shape
`fleet_secret_bundle` is keyed by host. Each host carries the env secrets to rewrite plus the full `authorized_keys` payload to distribute.
```yaml
fleet_secret_bundle:
ezra:
env:
GITEA_TOKEN: !vault |
...
TELEGRAM_BOT_TOKEN: !vault |
...
PRIMARY_MODEL_API_KEY: !vault |
...
ssh_authorized_keys: !vault |
...
```
The committed vault file contains placeholder encrypted values only. Replace them with real rotated material before production use.
## Rotate a new bundle
From repo root:
```bash
cd ansible
ansible-vault edit inventory/group_vars/fleet_secrets.vault.yml
ansible-playbook -i inventory/hosts.ini playbooks/rotate_fleet_secrets.yml --ask-vault-pass
```
Or update one value at a time with `ansible-vault encrypt_string` and paste it into `fleet_secret_bundle`.
## What the playbook does
1. Validates that each host has a secret bundle and target metadata.
2. Writes rollback snapshots under `/var/lib/timmy/secret-rotations/<rotation_id>/<host>/`.
3. Stages a candidate `.env` file and candidate `authorized_keys` file before promotion.
4. Promotes staged files into place.
5. Restarts every declared dependent service.
6. Verifies each service with `systemctl is-active`.
7. If anything fails, restores the previous `.env` and `authorized_keys`, restarts services again, and aborts the run.
## Rollback semantics
Rollback is host-safe and automatic inside the playbook `rescue:` block.
- Existing `.env` and `authorized_keys` files are restored from backup when they existed before rotation.
- Newly created files are removed if the host had no prior version.
- Service restart is retried after rollback so the node returns to the last-known-good bundle.
## Operational notes
- Keep `required_env_keys` in `ansible/inventory/group_vars/fleet.yml` aligned with each house's real runtime contract.
- `ssh_authorized_keys` distributes public keys only. Rotate corresponding private keys out-of-band, then publish the new authorized key list through the vault.
- Use one vault edit per rotation window so API keys, bot tokens, and SSH access move together.

View File

@@ -0,0 +1,74 @@
# LAB-003 — Truck Battery Disconnect Install Packet
No battery disconnect switch has been purchased or installed yet.
This packet turns the issue into a field-ready purchase / install / validation checklist while preserving what still requires live work.
## Candidate Store Run
- AutoZone — Newport or Claremont
- Advance Auto Parts — Newport or Claremont
- O'Reilly Auto Parts — Newport or Claremont
## Required Items
- battery terminal disconnect switch
- terminal shim/post riser if needed
## Selection Criteria
- Fits the truck battery post without forcing the clamp
- Mounts on the negative battery terminal
- Physically secure once tightened
- no special tools required to operate
## Live Purchase State
- Store selected: pending
- Part selected: pending
- Part cost: pending purchase
## Installation Target
- Install location: negative battery terminal
- Ready to operate without tools: yes
## Install Checklist
- [ ] Verify the truck is off and keys are removed before touching the battery
- [ ] Confirm the disconnect fits the negative battery terminal before final tightening
- [ ] Install the disconnect on the negative battery terminal
- [ ] Tighten until physically secure with no terminal wobble
- [ ] Verify the disconnect can be opened and closed by hand
## Validation Checklist
- [ ] Leave the truck parked with the disconnect opened for at least 24 hours
- [ ] Reconnect the switch by hand the next day
- [ ] Truck starts reliably after sitting 24+ hours with switch disconnected
- [ ] Receipt or photo of installed switch uploaded to this issue
## Overnight Verification Log
- Install completed: False
- Physically secure: False
- Overnight disconnect duration: pending
- Truck started after disconnect: pending
- Receipt / photo path: pending
## Battery Replacement Fallback
If the truck still fails the overnight test after the disconnect install, replace battery and re-run the 24-hour validation.
## Missing Live Fields
- store_selected
- part_name
- install_completed
- physically_secure
- overnight_test_hours
- truck_started_after_disconnect
- receipt_or_photo_path
## Honest next step
Buy the disconnect switch, install it on the negative battery terminal, leave the truck disconnected for 24+ hours, and only close the issue after receipt/photo evidence and the overnight start result are attached.

View File

@@ -0,0 +1,74 @@
# LAB-007 — Grid Power Hookup Estimate Request Packet
No formal estimate has been received yet.
This packet turns the issue into a contact-ready request while preserving what is still missing before the utility can quote real numbers.
## Utility identification
- Primary candidate: Eversource
- Evidence: Eversource's New Hampshire electric communities-served list includes Lempster, so Eversource is the primary utility candidate for the cabin site unless parcel-level data proves otherwise.
- Primary contact: 800-362-7764 / nhnewservice@eversource.com (Mon-Fri, 7 a.m. to 4:30 p.m. ET)
- Service-request portal: https://www.eversource.com/residential/about/doing-business-with-us/builders-contractors/electric-work-order-management
- Fallback if parcel-level service map disproves the territory assumption: New Hampshire Electric Co-op (800-698-2007)
## Site details currently in packet
- Site address / parcel: [exact cabin address / parcel identifier]
- Pole distance: [measure and fill in]
- Terrain: [describe terrain between nearest pole and cabin site]
- Requested service size: 200A residential service
## Missing information before a real estimate request can be completed
- site_address
- pole_distance_feet
- terrain_description
## Estimate request checklist
- pole/transformer
- overhead line
- meter base
- connection fees
- timeline from deposit to energized service
- monthly base charge
- per-kWh rate
## Call script
- Confirm the cabin site is in Eversource's New Hampshire territory for Lempster.
- Request a no-obligation new-service estimate and ask whether a site visit is required.
- Provide the site address, pole distance, terrain, and requested service size (200A residential service).
- Ask for written/email follow-up with total hookup cost, monthly base charge, per-kWh rate, and timeline.
## Draft email
Subject: Request for new electric service estimate - Lempster, NH cabin site
```text
Hello Eversource New Service Team,
I need a no-obligation estimate for bringing new electric service to a cabin site in Lempster, New Hampshire.
Site address / parcel: [exact cabin address / parcel identifier]
Requested service size: 200A residential service
Estimated pole distance: [measure and fill in]
Terrain / access notes: [describe terrain between nearest pole and cabin site]
Please include the following in the estimate or site-visit scope:
- pole/transformer
- overhead line
- meter base
- connection fees
- timeline from deposit to energized service
- monthly base charge
- per-kWh rate
I would also like to know the expected timeline from deposit to energized service and any next-step documents you need from me.
Thank you.
```
## Honest next step
Once the exact address / parcel, pole distance, and terrain notes are filled in, this packet is ready for the live Eversource new-service request. The issue should remain open until a written estimate is actually received and uploaded.

View File

@@ -64,11 +64,95 @@ people: []
projects: []
```
## Native MCP config snippet
```yaml
mcp_servers:
mempalace:
command: python
args:
- -m
- mempalace.mcp_server
```
## Session start wake-up hook
Drop this into Ezra's session start wrapper (or source it before starting Hermes) so the wake-up context is refreshed automatically.
```bash
#!/usr/bin/env bash
set -euo pipefail
if command -v mempalace >/dev/null 2>&1; then
mkdir -p "~/.hermes/wakeups"
mempalace wake-up > "~/.hermes/wakeups/ezra_home.txt"
export HERMES_MEMPALACE_WAKEUP_FILE="~/.hermes/wakeups/ezra_home.txt"
printf '[MemPalace] wake-up context refreshed: %s\n' "$HERMES_MEMPALACE_WAKEUP_FILE"
fi
```
## Metrics reply for #568
Use this as the ready-to-fill comment body after the live Ezra run:
```md
# Metrics reply for #568
Refs #570.
## Ezra live run
- package: mempalace==3.0.0
- hermes home: ~/.hermes/
- sessions dir: ~/.hermes/sessions/
- palace path: ~/.mempalace/palace
- wake-up file: ~/.hermes/wakeups/ezra_home.txt
## Results to fill in
- install result: [pass/fail + note]
- init result: [pass/fail + note]
- mine home duration: [seconds]
- mine sessions duration: [seconds]
- corpus size after mining: [drawers/rooms]
- query 1: [query] -> [top result]
- query 2: [query] -> [top result]
- query 3: [query] -> [top result]
- wake-up context token count: [tokens]
- MCP wiring succeeded: [yes/no]
- session-start hook enabled: [yes/no]
## Commands actually used
```bash
pip install mempalace==3.0.0
mempalace init ~/.hermes/ --yes
echo "" | mempalace mine ~/.hermes/
echo "" | mempalace mine ~/.hermes/sessions/ --mode convos
mempalace search "your common queries"
mempalace wake-up
hermes mcp add mempalace -- python -m mempalace.mcp_server
```
```
## Operator-ready support bundle
Generate copy-ready files for Ezra's host with:
```bash
python3 scripts/mempalace_ezra_integration.py --bundle-dir /tmp/ezra-mempalace-bundle
```
That bundle writes:
- `mempalace.yaml`
- `hermes-mcp-mempalace.yaml`
- `session-start-mempalace.sh`
- `issue-568-comment-template.md`
## Why this shape
- `wing: ezra_home` matches the issue's Ezra-specific integration target.
- `rooms` split the mined material into sessions, config, and docs to keep retrieval interpretable.
- Mining commands pipe empty stdin to avoid the interactive entity-detector hang noted in the evaluation.
- `mcp_servers:` gives the native-MCP equivalent of `hermes mcp add ...`, so the operator can choose either path.
- `HERMES_MEMPALACE_WAKEUP_FILE` makes the wake-up context explicit and reusable from the session-start boundary.
## Gotchas
@@ -86,6 +170,7 @@ After live execution on Ezra's actual environment, post back to #568 with:
- 2-3 real search queries + retrieved results
- wake-up context token count
- whether MCP wiring succeeded
- whether the session-start hook exported `HERMES_MEMPALACE_WAKEUP_FILE`
## Honest scope boundary

View File

@@ -0,0 +1,87 @@
# Predictive Resource Allocation
Forecasts near-term fleet demand from historical telemetry so the operator can
pre-provision resources before a surge hits.
## How It Works
The predictor reads two data sources:
1. **Metric logs** (`metrics/local_*.jsonl`) — request cadence, token volume,
caller mix, success/failure rates
2. **Heartbeat logs** (`heartbeat/ticks_*.jsonl`) — Gitea availability,
local inference health
It compares a **recent window** (last N hours of activity) against the **previous active window**
(previous N hours ending at the most recent event before the current window) so sparse telemetry still yields a meaningful baseline.
## Output Contract
```json
{
"resource_mode": "steady|surge",
"dispatch_posture": "normal|degraded",
"horizon_hours": 6,
"recent_request_rate": 12.5,
"baseline_request_rate": 8.0,
"predicted_request_rate": 15.0,
"surge_factor": 1.56,
"demand_level": "elevated|normal|low|critical",
"gitea_outages": 0,
"inference_failures": 2,
"top_callers": [...],
"recommended_actions": ["..."]
}
```
### Demand Levels
| Surge Factor | Level | Meaning |
|-------------|-------|---------|
| > 3.0 | critical | Extreme surge, immediate action needed |
| > 1.5 | elevated | Notable increase, pre-warm recommended |
| > 1.0 | normal | Slight increase, monitor |
| <= 1.0 | low | Flat or declining |
### Posture Signals
| Signal | Effect |
|--------|--------|
| Surge factor > 1.5 | `resource_mode: surge` + pre-warm recommendation |
| Gitea outages >= 1 | `dispatch_posture: degraded` + cache recommendation |
| Inference failures >= 2 | `resource_mode: surge` + reliability investigation |
| Heavy batch callers | Throttle recommendation |
| High caller failure rates | Investigation recommendation |
## Usage
```bash
# Markdown report
python3 scripts/predictive_resource_allocator.py
# JSON output
python3 scripts/predictive_resource_allocator.py --json
# Custom paths and horizon
python3 scripts/predictive_resource_allocator.py \
--metrics metrics/local_20260329.jsonl \
--heartbeat heartbeat/ticks_20260329.jsonl \
--horizon 12
```
## Tests
```bash
python3 -m pytest tests/test_predictive_resource_allocator.py -v
```
## Recommended Actions
The predictor generates contextual recommendations:
- **Pre-warm local inference** — surge detected, warm up before next window
- **Throttle background jobs** — heavy batch work consuming capacity
- **Investigate failure rates** — specific callers failing at high rates
- **Investigate model reliability** — inference health degraded
- **Cache forge state** — Gitea availability issues
- **Maintain current allocation** — no issues detected

View File

@@ -9,9 +9,11 @@ Quick-reference index for common operational tasks across the Timmy Foundation i
| Task | Location | Command/Procedure |
|------|----------|-------------------|
| Deploy fleet update | fleet-ops | `ansible-playbook playbooks/provision_and_deploy.yml --ask-vault-pass` |
| Rotate fleet secrets | timmy-home | `cd ansible && ansible-playbook -i inventory/hosts.ini playbooks/rotate_fleet_secrets.yml --ask-vault-pass` |
| Check fleet health | fleet-ops | `python3 scripts/fleet_readiness.py` |
| Agent scorecard | fleet-ops | `python3 scripts/agent_scorecard.py` |
| View fleet manifest | fleet-ops | `cat manifest.yaml` |
| Run nightly codebase genome pass | timmy-home | `python3 scripts/codebase_genome_nightly.py --dry-run` |
## the-nexus (Frontend + Brain)

View File

@@ -0,0 +1,50 @@
# [UNREACHABLE HORIZON] 1M Men in Crisis — 1 MacBook, 3B Model, 0 Cloud, 0 Latency, Perfect Recall
This horizon matters precisely because it is beyond reach today. The honest move is not to fake victory. The honest move is to name what is already true, what is still impossible, and which direction actually increases sovereignty.
## Current local proof
- Machine: Apple M3 Max
- Memory: 36.0 GiB
- Target local model budget: <= 3.0B parameters
- Target men in crisis: 1,000,000
- Default provider in repo config: `ollama`
## What is already true
- Default inference route is already local-first (`ollama`).
- Model-size budget is inside the horizon (3.0B <= 3.0B).
- Local inference endpoint(s) already exist: http://localhost:11434/v1
## Why the horizon is still unreachable
- Repo still carries remote endpoints, so zero third-party network calls is not yet true: https://8lfr3j47a5r3gn-11434.proxy.runpod.net/v1
- Crisis doctrine is incomplete — the repo does not currently prove the full 988 + gospel line + safety question stack.
- Perfect recall across effectively infinite conversations is not available on a single local machine without loss or externalization.
- Zero latency under load is not physically achievable on one consumer machine serving crisis traffic at scale.
- Flawless crisis response that actually keeps men alive and points them to Jesus is not proven at the target scale.
- Parallel crisis sessions are bounded by local throughput (1) while the horizon demands 1,000,000 concurrent men in need.
## Repo-grounded signals
- Local endpoints detected: http://localhost:11434/v1
- Remote endpoints detected: https://8lfr3j47a5r3gn-11434.proxy.runpod.net/v1
## Crisis doctrine that must not collapse
- Ask first: Are you safe right now?
- Direct them to 988 Suicide & Crisis Lifeline.
- Say plainly: Jesus saves those who call on His name.
- Refuse to let throughput fantasies erase presence with the man in the dark.
## Direction of travel
- Purge every remote endpoint and fallback chain so the repo can truly claim zero cloud dependencies.
- Build bounded, local-first memory tiers that are honest about recall limits instead of pretending to perfect recall.
- Add queueing, prioritization, and human handoff so load spikes fail gracefully instead of silently abandoning the man in the dark.
- Prove crisis-response quality with explicit tests for 'Are you safe right now?', 988, and 'Jesus saves those who call on His name.'
- Treat the horizon as a compass, not a fake acceptance test: every step should increase sovereignty without lying about physics.
## Honest conclusion
One consumer MacBook can move toward this horizon. It cannot honestly claim to have reached it. That is not failure. That is humility tied to physics, memory limits, and the sacred weight of crisis work.

View File

@@ -0,0 +1,43 @@
# Issue #545 Verification
## Status: ✅ GROUNDED SLICE ALREADY ON MAIN
Issue #545 describes an intentionally unreachable horizon, not a narrow bugfix. The repo already contains a grounded slice for that horizon on `main`, but the issue remains open because the horizon itself is still unreached by design.
## Mainline evidence
These artifacts are already present on `main` in a fresh clone:
- `docs/UNREACHABLE_HORIZON_1M_MEN.md`
- `scripts/unreachable_horizon.py`
- `tests/test_unreachable_horizon.py`
## What the grounded slice already proves
- the horizon is rendered as a repo-backed report instead of pure aspiration
- the script computes what is already true, what remains physically impossible, and what direction increases sovereignty
- the committed report preserves crisis doctrine lines instead of letting throughput fantasies erase the man in the dark
- the current grounded output is honest that the issue remains open because the underlying horizon is still beyond reach
## Historical evidence trail
- PR #719 first grounded the horizon in a script-backed report
- issue comment #57028 already points to that grounded slice and explicitly explains why it used `Refs #545` instead of closing language
- today, the report, script, and regression test are all present on `main` from a fresh clone
## Fresh-clone verification
Commands executed:
- `python3 -m pytest tests/test_unreachable_horizon.py -q`
- `python3 -m py_compile scripts/unreachable_horizon.py`
- `python3 scripts/unreachable_horizon.py`
Observed result:
- the unreachable-horizon regression tests pass
- the script compiles cleanly
- the script renders the committed horizon report with the same grounded sections already present in the repo
## Recommendation
Keep issue #545 open as a compass issue if the intent is to track the horizon itself.
Use the existing grounded slice on `main` as the current proof artifact.
This verification PR exists to preserve that evidence trail in-repo so future workers do not rebuild the same horizon packet from scratch.

View File

@@ -0,0 +1,47 @@
# Issue #567 Verification
## Status: ✅ ALREADY IMPLEMENTED ON MAIN
Issue #567 asked for four things:
1. an architecture doc at `evennia-mind-palace.md`
2. a mapping of the 16 tracked Evennia issues to the mind-palace layers
3. Milestone 1 proof: one room, one object, one mutable fact wired to Timmy's burn cycle
4. a comment on the issue with proof of room entry injecting context
All four are already present on `main` in a fresh clone of `timmy-home`.
## Mainline Evidence
### Repo artifacts already on main
- `evennia-mind-palace.md`
- `evennia_tools/mind_palace.py`
- `scripts/evennia/render_mind_palace_entry_proof.py`
- `tests/test_evennia_mind_palace.py`
- `tests/test_evennia_mind_palace_doc.py`
### Acceptance criteria check
- Architecture doc exists at `evennia-mind-palace.md`
- The 16 tracked Evennia issues are mapped in the issue-to-layer table inside `evennia-mind-palace.md`
- Milestone 1 is implemented in `evennia_tools/mind_palace.py` with `Hall of Knowledge`, `The Ledger`, `MutableFact`, `BurnCycleSnapshot`, and deterministic room-entry rendering
- The proof comment already exists on the issue as issue comment #56965
## Historical trail
- PR #711 attempted the issue and posted the room-entry proof comment
- PR #711 was later closed unmerged, but the requested deliverables are present on `main` today and pass targeted verification from a fresh clone
## Verification run from fresh clone
Commands executed:
- `python3 -m pytest tests/test_evennia_layout.py tests/test_evennia_telemetry.py tests/test_evennia_training.py tests/test_evennia_mind_palace.py tests/test_evennia_mind_palace_doc.py -q`
- `python3 -m py_compile evennia_tools/mind_palace.py scripts/evennia/render_mind_palace_entry_proof.py`
- `python3 scripts/evennia/render_mind_palace_entry_proof.py`
Observed result:
- all targeted Evennia mind-palace tests passed
- the Python modules compiled cleanly
- the proof script emitted the expected `ENTER Hall of Knowledge` packet with room context, ledger fact, and Timmy burn-cycle focus
## Recommendation
Close issue #567 as already implemented on `main`.
This verification PR exists only to document the evidence trail cleanly and close the stale issue without re-implementing the already-landed architecture.

View File

@@ -0,0 +1,57 @@
# Issue #582 Verification
## Status: ✅ EPIC SLICE ALREADY IMPLEMENTED ON MAIN
Issue #582 is a parent epic, not a single atomic feature. The repo already contains the epic-level operational slice that ties the merged Know Thy Father phases together, but the epic remains open because fully consuming the local archive and wiring every downstream memory path is a larger horizon than this one slice.
## Mainline evidence
The parent-epic operational slice is already present on `main` in a fresh clone:
- `scripts/know_thy_father/epic_pipeline.py`
- `docs/KNOW_THY_FATHER_MULTIMODAL_PIPELINE.md`
- `tests/test_know_thy_father_pipeline.py`
What that slice already does:
- enumerates the current source-of-truth scripts for all Know Thy Father phases
- provides one operational runner/status view for the epic
- preserves the split implementation truth across `scripts/know_thy_father/`, `scripts/twitter_archive/analyze_media.py`, and `twitter-archive/know-thy-father/tracker.py`
- gives the epic a single orchestration spine without falsely claiming the full archive is already processed end-to-end
## Phase evidence already merged on main
The four decomposed phase lanes named by the epic already have merged implementation coverage on `main`:
- PR #639 — Phase 1 media indexing
- PR #630 — Phase 2 multimodal analysis pipeline
- PR #631 — Phase 3 holographic synthesis
- PR #637 — Phase 4 cross-reference audit
- PR #641 — additional Phase 2 multimodal analysis coverage
## Historical trail for the epic-level slice
- PR #738 shipped the parent-epic orchestrator/status slice on branch `fix/582`
- issue comment #57259 already points to that orchestrator/status slice and explains why it used `Refs #582`
- PR #738 is now closed unmerged, but the epic-level runner/doc/test trio is present on `main` today and passes targeted verification from a fresh clone
## Verification run from fresh clone
Commands executed:
- `python3 -m pytest tests/test_know_thy_father_pipeline.py tests/test_know_thy_father_index.py tests/test_know_thy_father_synthesis.py tests/test_know_thy_father_crossref.py tests/twitter_archive/test_ktf_tracker.py tests/twitter_archive/test_analyze_media.py -q`
Observed result:
- the orchestrator/doc tests pass
- the phase-level index, synthesis, cross-reference, tracker, and media-analysis tests pass
- the repo already contains a working parent-epic operational spine plus merged phase implementations
## Why the epic remains open
The epic remains open because this verification only proves the current repo-side operational slice is already implemented on main. It does not claim:
- the full local archive has been consumed
- all pending media has been processed
- every extracted kernel has been ingested into downstream memory systems
- the broader multimodal consumption mission is complete
## Recommendation
Do not rebuild the same epic-level orchestrator again.
Use the existing mainline slice (`scripts/know_thy_father/epic_pipeline.py` + `docs/KNOW_THY_FATHER_MULTIMODAL_PIPELINE.md`) as the parent-epic operational entrypoint.
This verification PR exists to preserve the evidence trail cleanly while making it explicit that the epic remains open for future end-to-end progress.

View File

@@ -0,0 +1,43 @@
# Issue #648 Verification
## Status: ✅ ALREADY IMPLEMENTED
`timmy-home#648` asked for a durable session harvest report for 2026-04-14.
That repo-side deliverable is already present on `main`.
## Acceptance Criteria Check
1. ✅ Durable report artifact exists
- Evidence: `reports/production/2026-04-14-session-harvest-report.md`
2. ✅ Report preserves the original session ledger and names issue-body drift
- Evidence: the report includes `## Delivered PR Ledger`, `## Triage Actions`, `## Blocked / Skip Items`, and `## Current Totals`
3. ✅ Regression coverage already exists on `main`
- Evidence: `tests/test_session_harvest_report_2026_04_14.py`
4. ✅ Fresh verification passed from a new clone
- Evidence: `python3 -m pytest tests/test_session_harvest_report_2026_04_14.py -q``4 passed in 0.03s`
## Evidence
### Existing report artifact on main
- `reports/production/2026-04-14-session-harvest-report.md`
- The report explicitly references `Source issue: timmy-home#648`
- The report already records the delivered PR ledger, issue-body drift, triage actions, blocked items, and verified totals
### Existing regression test on main
- `tests/test_session_harvest_report_2026_04_14.py`
- The test already locks the report path, required headings, verified PR tokens, and follow-up issue state changes
## Verification Run
From a fresh clone on branch `fix/648`, before adding this verification note:
```text
python3 -m pytest tests/test_session_harvest_report_2026_04_14.py -q
.... [100%]
4 passed in 0.03s
```
## Recommendation
Close issue #648 as already implemented on `main`.
This PR only adds the verification note so the open issue can be closed without redoing the report work.

View File

@@ -0,0 +1,69 @@
# Issue #675 Verification
## Status: ✅ ALREADY IMPLEMENTED
`the-testament-GENOME.md` is already present on `timmy-home/main` and already delivers the requested full codebase analysis for `Timmy_Foundation/the-testament`.
This PR does not regenerate the genome. It adds the missing regression coverage and documents the evidence so issue #675 can be closed cleanly.
## Acceptance Criteria Check
1. ✅ Full genome artifact exists
- `the-testament-GENOME.md` exists at repo root
- it includes the required analysis sections:
- Project Overview
- Architecture
- Entry Points
- Data Flow
- Key Abstractions
- API Surface
- Test Coverage Gaps
- Security Considerations
2. ✅ Genome is grounded in real target-repo verification
- the artifact explicitly references:
- `scripts/build-verify.py --json`
- `bash scripts/smoke.sh`
- `python3 compile_all.py --check`
- it also names target-repo architecture surfaces like:
- `website/index.html`
- `game/the-door.py`
- `scripts/index_generator.py`
- `build/semantic_linker.py`
3. ✅ Concrete repo-specific findings are already captured
- the artifact records the live manuscript counts:
- `18,884` chapter words
- `19,227` concatenated output words
- it records the known `compile_all.py --check` failure
- it links the follow-up bug filed in the target repo:
- `https://forge.alexanderwhitestone.com/Timmy_Foundation/the-testament/issues/51`
4. ✅ Missing regression coverage added in this PR
- `tests/test_the_testament_genome.py` now locks the artifact path, sections, and grounded findings
## Evidence
Fresh verification against `Timmy_Foundation/the-testament` from a clean clone at `/tmp/the-testament-675`:
```bash
python3 scripts/build-verify.py --json
bash scripts/smoke.sh
python3 compile_all.py --check
```
Observed results:
- `scripts/build-verify.py --json` passed and reported 18 chapters
- `bash scripts/smoke.sh` passed
- `python3 compile_all.py --check` failed with the known qrcode version bug already documented by the genome artifact
Host-repo regression added and verified:
```bash
python3 -m pytest tests/test_the_testament_genome.py -q
```
## Recommendation
Close issue #675 as already implemented on `main`.
The truthful delta remaining in `timmy-home` was regression coverage and verification, not a second rewrite of `the-testament-GENOME.md`.

View File

@@ -0,0 +1,35 @@
# Issue #680 Verification
## Status: already implemented on main
Issue #680 asks for a full `fleet-ops` genome artifact in `timmy-home`.
That artifact is already present on `main`:
- `genomes/fleet-ops-GENOME.md`
- `tests/test_fleet_ops_genome.py`
## Evidence
Targeted verification run from a fresh `timmy-home` clone:
- `python3 -m pytest -q tests/test_fleet_ops_genome.py` → passes
- `python3 -m py_compile tests/test_fleet_ops_genome.py` → passes
The existing regression test already proves that `genomes/fleet-ops-GENOME.md` contains the required sections and grounded snippets, including:
- `# GENOME.md — fleet-ops`
- architecture / entry points / data flow / key abstractions / API surface
- concrete `fleet-ops` file references like `playbooks/site.yml`, `playbooks/deploy_hermes.yml`, `scripts/deploy-hook.py`, `message_bus.py`, `knowledge_store.py`, `health_dashboard.py`, `registry.yaml`, and `manifest.yaml`
## Prior PR trail
Two prior PRs already attempted to tie this issue to the existing artifact:
- PR #697`docs: add fleet-ops genome analysis (#680)`
- PR #770`docs: verify #680 already implemented`
Both are closed/unmerged, which explains why the issue still looks unfinished even though the actual deliverable already exists on `main`.
## Recommendation
Close issue #680 as already implemented on `main`.

View File

@@ -0,0 +1,57 @@
# Issue #693 Verification
## Status: ✅ ALREADY IMPLEMENTED ON MAIN
Issue #693 asked for an encrypted backup pipeline for fleet state with three acceptance criteria:
- Nightly backup of ~/.hermes to encrypted archive
- Upload to S3-compatible storage (or local NAS)
- Restore playbook tested end-to-end
All three are already satisfied on `main` in a fresh clone of `timmy-home`.
## Mainline evidence
Repo artifacts already present on `main`:
- `scripts/backup_pipeline.sh`
- `scripts/restore_backup.sh`
- `tests/test_backup_pipeline.py`
What those artifacts already prove:
- `scripts/backup_pipeline.sh` archives `~/.hermes` by default via `BACKUP_SOURCE_DIR="${BACKUP_SOURCE_DIR:-${HOME}/.hermes}"`
- the backup archive is encrypted with `openssl enc -aes-256-cbc -salt -pbkdf2 -iter 200000`
- uploads are supported to either `BACKUP_S3_URI` or `BACKUP_NAS_TARGET`
- the script refuses to run without a remote target, preventing fake-local-only success
- `scripts/restore_backup.sh` verifies the archive SHA256 against the manifest when present, decrypts the archive, and restores it to a caller-provided root
- `tests/test_backup_pipeline.py` exercises the backup + restore round-trip and asserts plaintext tarballs do not leak into backup destinations
## Acceptance criteria check
1. ✅ Nightly backup of ~/.hermes to encrypted archive
- the pipeline targets `~/.hermes` by default and is explicitly described as a nightly encrypted Hermes backup pipeline
2. ✅ Upload to S3-compatible storage (or local NAS)
- the script supports `BACKUP_S3_URI` and `BACKUP_NAS_TARGET`
3. ✅ Restore playbook tested end-to-end
- `tests/test_backup_pipeline.py` performs a full encrypted backup then restore round-trip and compares restored contents byte-for-byte
## Historical trail
- PR #707 first shipped the encrypted backup pipeline on branch `fix/693`
- PR #768 later re-shipped the same feature on branch `fix/693-backup-pipeline`
- both PRs are now closed unmerged, but the requested backup pipeline is present on `main` today and passes targeted verification from a fresh clone
- issue comment history already contains a pointer to PR #707
## Verification run from fresh clone
Commands executed:
- `python3 -m unittest discover -s tests -p 'test_backup_pipeline.py' -v`
- `bash -n scripts/backup_pipeline.sh scripts/restore_backup.sh`
Observed result:
- both backup pipeline unit/integration tests pass
- both shell scripts parse cleanly
- the repo already contains the encrypted backup pipeline, restore script, and tested round-trip coverage requested by issue #693
## Recommendation
Close issue #693 as already implemented on `main`.
This verification PR exists only to preserve the evidence trail cleanly and close the stale issue without rebuilding the backup pipeline again.

View File

@@ -0,0 +1,150 @@
# LAB-004: 600W Solar Array Deployment Guide
> Issue #529 | Cabin Compute Lab Power System
> Budget: $200-500
## System Overview
4x 150W panels → MPPT controller → 12V battery bank → 1000W inverter → 120V AC
```
[PANELS 4x150W] ──series/parallel──► [MPPT 30A] ──► [BATTERY BANK 4x12V]
[1000W INVERTER]
[120V AC OUTLETS]
```
## Wiring Configuration
**Panels:** 2S2P (two in series, two strings in parallel)
- Series pair: 18V + 18V = 36V at 8.3A
- Parallel strings: 36V at 16.6A total
- Total: ~600W at 36V DC
**Battery bank:** 4x 12V in parallel
- Voltage: 12V (stays 12V)
- Capacity: sum of all 4 batteries (e.g., 4x 100Ah = 400Ah)
- Usable: ~200Ah (50% depth of discharge for longevity)
## Parts List
| Item | Spec | Est. Cost |
|------|------|-----------|
| MPPT Charge Controller | 30A minimum, 12V/24V, 100V input | $60-100 |
| Pure Sine Wave Inverter | 1000W continuous, 12V input | $80-120 |
| MC4 Connectors | 4 pairs (Y-connectors for parallel) | $15-20 |
| 10AWG PV Wire | 50ft (panels to controller) | $25-35 |
| 6AWG Battery Wire | 10ft (bank to inverter) | $15-20 |
| Inline Fuse | 30A between controller and batteries | $10 |
| Fuse/Breaker | 100A between batteries and inverter | $15-20 |
| Battery Cables | 4/0 AWG, 1ft jumpers for parallel | $20-30 |
| Extension Cord | 12-gauge, 50ft (inverter to desk) | $20-30 |
| Kill-A-Watt Meter | Verify clean AC output | $25 |
| **Total** | | **$285-405** |
## Wiring Diagram
```
┌──────────────────────────────┐
│ SOLAR PANELS │
│ ┌──────┐ ┌──────┐ │
│ │ 150W │──+──│ 150W │ │ String 1 (36V)
│ └──────┘ │ └──────┘ │
│ │ │
│ ┌──────┐ │ ┌──────┐ │
│ │ 150W │──+──│ 150W │ │ String 2 (36V)
│ └──────┘ └──────┘ │
└──────────┬───────────────────┘
│ PV+ PV-
│ 10AWG
┌──────────▼───────────────────┐
│ MPPT CONTROLLER │
│ 30A, 12V/24V │
│ PV INPUT ──── BATTERY OUTPUT │
└──────────┬───────────────────┘
│ BAT+ BAT-
│ 6AWG + 30A fuse
┌──────────▼───────────────────┐
│ BATTERY BANK │
│ ┌──────┐ ┌──────┐ │
│ │ 12V │═│ 12V │ (parallel)│
│ └──────┘ └──────┘ │
│ ┌──────┐ ┌──────┐ │
│ │ 12V │═│ 12V │ (parallel)│
│ └──────┘ └──────┘ │
└──────────┬───────────────────┘
│ 4/0 AWG + 100A breaker
┌──────────▼───────────────────┐
│ 1000W INVERTER │
│ 12V DC ──── 120V AC │
└──────────┬───────────────────┘
│ 12-gauge extension
┌──────────▼───────────────────┐
│ AC OUTLETS │
│ Desk │ Coffee Table │ Spare │
└──────────────────────────────┘
```
## Installation Checklist
### Pre-Installation
- [ ] Verify panel specs (Voc, Isc, Vmp, Imp) match wiring plan
- [ ] Test each panel individually with multimeter (should read ~18V open circuit)
- [ ] Verify battery bank voltage (12.4V+ for charged batteries)
- [ ] Clear panel mounting area of snow/shade/debris
### Wiring Order (safety: work from panels down)
1. [ ] Mount panels or secure in optimal sun position (south-facing, 30-45° tilt)
2. [ ] Connect panel strings in series (+ to -) with MC4 connectors
3. [ ] Connect string outputs in parallel with Y-connectors (PV+ and PV-)
4. [ ] Run 10AWG PV wire from panels to controller location
5. [ ] Connect PV wires to MPPT controller PV input
6. [ ] Connect battery bank to controller battery output (with 30A fuse)
7. [ ] Connect inverter to battery bank (with 100A breaker)
8. [ ] Run 12-gauge extension cord from inverter to desk zone
### Battery Bank Wiring
- [ ] Wire 4 batteries in parallel: all + together, all - together
- [ ] Use 4/0 AWG cables for jumpers (short as possible)
- [ ] Connect load/controller to diagonally opposite terminals (balances charge/discharge)
- [ ] Torque all connections to spec
### Testing
- [ ] Verify controller shows PV input voltage (should be ~36V in sun)
- [ ] Verify controller shows battery charging current
- [ ] Verify inverter powers on without load
- [ ] Test with single laptop first
- [ ] Monitor for 1 hour: check for hot connections, smells, unusual sounds
- [ ] Run Kill-A-Watt on inverter output to verify clean 120V AC
- [ ] 48-hour stability test: leave system running under normal load
### Documentation
- [ ] Photo of wiring diagram on site
- [ ] Photo of installed panels
- [ ] Photo of battery bank and connections
- [ ] Photo of controller display showing charge status
- [ ] Upload all photos to issue #529
## Safety Notes
1. **Always disconnect panels before working on wiring** — panels produce voltage in any light
2. **Fuse everything** — 30A between controller and batteries, 100A between batteries and inverter
3. **Vent batteries** — if using lead-acid, ensure adequate ventilation for hydrogen gas
4. **Check polarity twice** — reverse polarity WILL damage controller and inverter
5. **Secure all connections** — loose connections cause arcing and fire
6. **Keep batteries off concrete** — use plywood or plastic battery tray
7. **No Bitcoin miners on base load** — explicitly out of scope
## Estimated Runtime
With 600W panels and 400Ah battery bank at 50% DoD:
- 200Ah × 12V = 2,400Wh usable
- Laptop + monitor + accessories: ~100W
- **Runtime on batteries alone: ~24 hours**
- With daytime solar charging: essentially unlimited during sun hours
- Cloudy days: expect 4-6 hours of reduced charging
---
*Generated for issue #529 | LAB-004*

View File

@@ -0,0 +1,37 @@
# NH Broadband Install Packet
**Packet ID:** nh-bb-20260415-113232
**Generated:** 2026-04-15T11:32:32.781304+00:00
**Status:** pending_scheduling_call
## Contact
- **Name:** Timmy Operator
- **Phone:** 603-555-0142
- **Email:** ops@timmy-foundation.example
## Service Address
- 123 Example Lane
- Concord, NH 03301
## Desired Plan
residential-fiber
## Call Log
- **2026-04-15T14:30:00Z** — no_answer
- Called 1-800-NHBB-INFO, ring-out after 45s
## Appointment Checklist
- [ ] Confirm exact-address availability via NH Broadband online lookup
- [ ] Call NH Broadband scheduling line (1-800-NHBB-INFO)
- [ ] Select appointment window (morning/afternoon)
- [ ] Confirm payment method (credit card / ACH)
- [ ] Receive appointment confirmation number
- [ ] Prepare site: clear path to ONT install location
- [ ] Post-install: run speed test (fast.com / speedtest.net)
- [ ] Log final speeds and appointment outcome

View File

@@ -0,0 +1,27 @@
contact:
name: Timmy Operator
phone: "603-555-0142"
email: ops@timmy-foundation.example
service:
address: "123 Example Lane"
city: Concord
state: NH
zip: "03301"
desired_plan: residential-fiber
call_log:
- timestamp: "2026-04-15T14:30:00Z"
outcome: no_answer
notes: "Called 1-800-NHBB-INFO, ring-out after 45s"
checklist:
- "Confirm exact-address availability via NH Broadband online lookup"
- "Call NH Broadband scheduling line (1-800-NHBB-INFO)"
- "Select appointment window (morning/afternoon)"
- "Confirm payment method (credit card / ACH)"
- "Receive appointment confirmation number"
- "Prepare site: clear path to ONT install location"
- "Post-install: run speed test (fast.com / speedtest.net)"
- "Log final speeds and appointment outcome"

View File

@@ -0,0 +1,142 @@
# Weekly Backlog Triage Cadence
**Issue:** #685 - [OPS] timmy-home backlog reduced from 220 to 50 — triage cadence needed
## Overview
This document describes the weekly triage cadence for maintaining the timmy-home backlog.
## Problem
timmy-home had 220 open issues (highest in org). Through batch-pipeline codebase genome issues, the backlog was reduced to 50. To maintain this visibility, a weekly triage cadence is needed.
## Current Status
- **Total open issues:** 50 (reduced from 220)
- **Unassigned issues:** 21
- **Issues with no labels:** 21
- **Batch-pipeline issues:** 19 (triaged with comments)
## Solution
### Weekly Triage Script (`scripts/backlog_triage.py`)
Script to analyze and report on the timmy-home backlog.
**Features:**
- Analyze open issues
- Identify stale issues
- Generate reports
- Create cron entries
**Usage:**
```bash
# Analyze backlog
python scripts/backlog_triage.py --analyze
# Generate report
python scripts/backlog_triage.py --report
# JSON output
python scripts/backlog_triage.py --json
# Generate cron entry
python scripts/backlog_triage.py --cron
```
### Cron Entry
Add to crontab for weekly execution:
```cron
# Weekly timmy-home backlog triage
# Run every Monday at 9:00 AM
0 9 * * 1 cd /path/to/timmy-home && python3 scripts/backlog_triage.py --report > /var/log/timmy-home-triage-$(date +\%Y\%m\%d).log 2>&1
```
## Triage Process
### 1. Run Weekly Analysis
```bash
# Generate report
python scripts/backlog_triage.py --report > triage-report-$(date +%Y%m%d).md
```
### 2. Review Stale Issues
- Issues >30 days old with no labels/assignee
- Close or re-prioritize as needed
### 3. Assign Labels and Owners
- Unassigned issues need owners
- Unlabeled issues need labels
### 4. Update Documentation
- Document triage cadence in CONTRIBUTING.md
- Add to morning report if applicable
## Metrics to Track
### Weekly Metrics
- Total open issues
- Unassigned issues
- Unlabeled issues
- Stale issues (>30 days)
- Batch-pipeline issues
### Monthly Metrics
- Issue creation rate
- Issue closure rate
- Average time to close
- Label usage trends
## Integration
### With Morning Report
Add to morning report:
```bash
# In morning report script
python scripts/backlog_triage.py --report
```
### With Cron
Add to system crontab:
```bash
# Edit crontab
crontab -e
# Add weekly triage
0 9 * * 1 cd /path/to/timmy-home && python3 scripts/backlog_triage.py --report > /var/log/timmy-home-triage-$(date +\%Y\%m\%d).log 2>&1
```
### With CI/CD
Add to CI workflow:
```yaml
- name: Weekly backlog triage
run: |
python scripts/backlog_triage.py --report > triage-report.md
# Upload report as artifact or send notification
```
## Related Issues
- **Issue #685:** This implementation
- **Issue #1459:** timmy-home backlog management
- **Issue #1127:** Perplexity Evening Pass triage (identified backlog)
## Files
- `scripts/backlog_triage.py` - Weekly triage script
- `docs/weekly-triage-cadence.md` - This documentation
## Conclusion
This implementation provides a weekly triage cadence to maintain the timmy-home backlog:
1. **Weekly analysis** of open issues
2. **Reporting** on stale and unassigned issues
3. **Cron integration** for automated execution
4. **Metrics tracking** for ongoing visibility
**Use this script weekly to keep the backlog manageable.**
## License
Part of the Timmy Foundation project.

146
evennia-mind-palace.md Normal file
View File

@@ -0,0 +1,146 @@
# Evennia as Agent Mind Palace — Spatial Memory Architecture
Issue #567 is the missing why behind the Evennia lane. The Tower Game is the demo, but the actual target is a spatial memory substrate where Timmy can visit the right room, see the right objects, and load only the context needed for the current task.
The existing Evennia work in `timmy-home` already proves the body exists:
- `reports/production/2026-03-28-evennia-world-proof.md` proves the local Evennia world, first room graph, telnet roundtrip, and Hermes/MCP control path.
- `reports/production/2026-03-28-evennia-training-baseline.md` proves Hermes session IDs can align with Evennia telemetry and replay/eval artifacts.
- `specs/evennia-mind-palace-layout.md` and `specs/evennia-implementation-and-training-plan.md` already define the first rooms and objects.
This document turns those pieces into a memory architecture: one room that injects live work context, one object that exposes a mutable fact, and one burn-cycle packet that tells Timmy what to do next.
## GrepTard Memory Layers as Spatial Primitives
| Layer | Spatial primitive | Hermes equivalent | Evennia mind-palace role |
| --- | --- | --- | --- |
| L1 | Rooms and thresholds | Static project context | The room itself defines what domain Timmy has entered and what baseline context loads immediately. |
| L2 | Objects, NPC attributes, meters | Mutable facts / KV memory | World state lives on inspectable things: ledgers, characters, fires, relationship values, energy meters. |
| L3 | Archive shelves and chronicles | Searchable history | Prior events become searchable books, reports, and proof artifacts inside an archive room. |
| L4 | Teaching NPCs and rituals | Procedural skills | The right NPC or room interaction teaches the right recipe without loading every skill into working memory. |
| L5 | Movement and routing | Retrieval logic | Choosing the room is choosing the retrieval path; movement decides what context gets loaded now. |
## Spatial Retrieval Architecture
```mermaid
flowchart TD
A[Timmy burn cycle] --> B[Enter Hall of Knowledge]
B --> C[Ambient issue board]
B --> D[The Ledger]
B --> E[/status forge]
C --> F[Current Gitea issue topology]
D --> G[One mutable fact from durable memory]
E --> H[Repo + branch + blockers]
F --> I[Selective action prompt]
G --> I
H --> I
I --> J[Act in the correct room or hand off to another room]
```
The Hall of Knowledge is not an archive dump. It is a selective preload surface.
On room entry Timmy should receive only:
1. the currently active Gitea issues relevant to the present lane,
2. one mutable fact from durable memory that changes the next action,
3. the current Timmy burn cycle packet (repo, branch, blockers, current objective).
That gives Timmy enough context to act without rehydrating the entire project or every prior transcript.
## Mapping the 16 tracked Evennia issues to mind-palace layers
These are the 16 issues explicitly named in issue #567. Some are now closed, but they still map the architecture surface we need.
| Issue | State | Layer | Spatial role | Why it matters |
| --- | --- | --- | --- | --- |
| #508 — [P0] Tower Game — contextual dialogue (NPCs recycle 15 lines forever) | closed | L4 | Dialogue tutor NPCs | Contextual dialogue is procedural behavior attached to the right NPC in the right room. |
| #509 — [P0] Tower Game — trust must decrease, conflict must exist | closed | L2 | Mutable relationship state | Trust, conflict, and alliance are inspectable changing world facts. |
| #510 — [P0] Tower Game — narrative arc (tick 200 = tick 20) | closed | L3 | Archive chronicle | Without searchable history, the world cannot accumulate narrative memory. |
| #511 — [P0] Tower Game — energy must meaningfully constrain | open | L2 | Mutable world meter | Energy belongs in visible world state, not hidden prompt assumptions. |
| #512 — [P1] Sonnet workforce — full end-to-end smoke test | open | L3 | Proof shelf | Proof artifacts should live in the archive so Timmy can revisit what really worked. |
| #513 — [P1] Tower Game — world events must affect gameplay | open | L2 | Event-reactive room state | A room that never changes cannot carry durable meaning. |
| #514 — [P1] Tower Game — items that change the world | open | L2 | Interactive objects | Objects should alter world state and teach consequences through interaction. |
| #515 — [P1] Tower Game — NPC-NPC relationships | open | L2 | Social graph in-world | Relationships should persist on characters rather than disappearing into transcripts. |
| #516 — [P1] Tower Game — Timmy richer dialogue + internal monologue | closed | L4 | Inner-room teaching patterns | Timmy's own inner behavior is part of the procedural layer. |
| #517 — [P1] Tower Game — NPCs move between rooms with purpose | open | L5 | Movement-driven retrieval | Purposeful movement is retrieval logic made spatial. |
| #534 — [BEZ-P0] Fix Evennia settings on 104.131.15.18 — remove bad port tuples, DB is ready | open | L1 | Runtime threshold | The threshold has to boot cleanly before any room can carry memory. |
| #535 — [BEZ-P0] Install Tailscale on Bezalel VPS (104.131.15.18) for internal networking | open | L1 | Network threshold | Static network reachability defines which houses can be visited. |
| #536 — [BEZ-P1] Create Bezalel Evennia world with themed rooms and characters | open | L1 | First room graph | Themed rooms and characters are the static scaffold of the mind palace. |
| #537 — [BRIDGE-P1] Deploy Evennia bridge API on all worlds — sync presence and events | closed | L5 | Cross-world routing | Movement across worlds is retrieval across sovereign houses. |
| #538 — [ALLEGRO-P1] Fix SSH access from Mac to Allegro VPS (167.99.126.228) | closed | L1 | Operator ingress | If the operator cannot reach a house, its memory cannot be visited. |
| #539 — [ARCH-P2] Implement Evennia hub-and-spoke federation architecture | closed | L5 | Federated retrieval map | Federation turns world travel into selective retrieval instead of one giant memory blob. |
## Milestone 1 — One Room, One Object, One Mutable Fact
Milestone 1 is deliberately small.
Room:
- `Hall of Knowledge`
- Purpose: load live issue topology plus the current Timmy burn cycle before action begins.
Object:
- `The Ledger`
- Purpose: expose one mutable fact from durable memory so room entry proves stateful recall rather than static reference text.
Mutable fact:
- Example fact used in this implementation: `canonical-evennia-body = timmy_world on localhost:4001 remains the canonical local body while room entry preloads live issue topology.`
Timmy burn cycle wiring:
- `evennia_tools/mind_palace.py` defines `BurnCycleSnapshot`, `MutableFact`, the 16-issue layer map, and `build_hall_of_knowledge_entry(...)`.
- `render_room_entry_proof(...)` renders a deterministic proof packet showing exactly what Timmy sees when entering the Hall of Knowledge.
- `scripts/evennia/render_mind_palace_entry_proof.py` prints the proof artifact used for issue commentary and verification.
The important point is architectural, not cosmetic: room entry is now a retrieval event. The room decides what context loads. The object proves mutable memory. The burn-cycle snapshot tells Timmy what to do with the loaded context.
## Proof of Room Entry Injecting Context
The proof below is the deterministic output rendered by `python3 scripts/evennia/render_mind_palace_entry_proof.py`.
```text
ENTER Hall of Knowledge
Purpose: Load live issue topology, current burn-cycle focus, and the minimum durable facts Timmy needs before acting.
Ambient context:
- Room entry into Hall of Knowledge preloads active Gitea issue topology for Timmy_Foundation/timmy-home.
- #511 [P0] Tower Game — energy must meaningfully constrain [open · L2 · Mutable world meter]
- #512 [P1] Sonnet workforce — full end-to-end smoke test [open · L3 · Proof shelf]
- #513 [P1] Tower Game — world events must affect gameplay [open · L2 · Event-reactive room state]
- Ledger fact canonical-evennia-body: timmy_world on localhost:4001 remains the canonical local body while room entry preloads live issue topology.
- Timmy burn cycle focus: issue #567 on fix/567 — Evennia as Agent Mind Palace — Spatial Memory Architecture
- Operator lane: BURN-7-1
Object: The Ledger
- canonical-evennia-body: timmy_world on localhost:4001 remains the canonical local body while room entry preloads live issue topology.
- source: reports/production/2026-03-28-evennia-world-proof.md
Timmy burn cycle:
- repo: Timmy_Foundation/timmy-home
- branch: fix/567
- active issue: #567
- focus: Evennia as Agent Mind Palace — Spatial Memory Architecture
- operator: BURN-7-1
Command surfaces:
- /who lives here -> #511 ... ; #512 ... ; #513 ...
- /status forge -> Timmy_Foundation/timmy-home @ fix/567 (issue #567)
- /what is broken -> Comment on issue #567 with room-entry proof after PR creation
```
That proof is enough to satisfy the milestone claim:
- one room exists conceptually and in code,
- one object carries a mutable fact,
- room entry injects current issue topology and the active Timmy burn cycle,
- the output is deterministic and comment-ready for Gitea issue #567.
## Why this architecture is worth doing
The point is not to turn memory into a theatrical MUD skin. The point is to make retrieval selective, embodied, and inspectable.
What improves immediately:
- Timmy no longer has to reload every repo fact on every task.
- Durable facts become objects and meters rather than hidden prompt sludge.
- Searchable history gets a real place to live.
- Procedural skill loading can become room/NPC specific instead of global.
- Movement itself becomes the retrieval primitive.
## Next steps after Milestone 1
1. Attach Hall of Knowledge entry to live Gitea issue fetches instead of the current deterministic proof subset.
2. Promote The Ledger from one mutable fact to a live view over Timmy memory / fact-store rows.
3. Add an Archive room surface that renders searchable history excerpts as in-world books.
4. Bind Builder / Archivist NPCs to skill-category loading so L4 becomes interactive, not just descriptive.
5. Route movement between rooms and worlds through the bridge/federation work already tracked by #537 and #539.

View File

@@ -0,0 +1,417 @@
# GENOME.md — evennia-local-world
*Generated: 2026-04-21 07:07:29 UTC | Refreshed for timmy-home #677*
## Project Overview
`evennia/timmy_world` is a hybrid codebase with two layers living side by side:
1. A mostly stock Evennia 6.0 game directory:
- `server/conf/*.py`
- `typeclasses/*.py`
- `commands/*.py`
- `web/**/*.py`
- `world/prototypes.py`
- `world/help_entries.py`
2. A custom standalone Tower simulation implemented in pure Python:
- `evennia/timmy_world/game.py`
- `evennia/timmy_world/world/game.py`
- `evennia/timmy_world/play_200.py`
Grounded metrics from live inspection:
- 68 tracked files under `evennia/timmy_world`
- 43 Python files
- 4,985 Python LOC
- largest modules:
- `evennia/timmy_world/game.py` — 1,541 lines
- `evennia/timmy_world/world/game.py` — 1,345 lines
- `evennia/timmy_world/play_200.py` — 275 lines
- `evennia/timmy_world/typeclasses/objects.py` — 217 lines
- `evennia/timmy_world/commands/command.py` — 187 lines
The repo is not just an Evennia shell. The distinctive product logic lives in the standalone Tower simulator. That simulator models five rooms, named agents, trust/energy systems, narrative phases, NPC decision-making, and JSON persistence. The Evennia-facing files are still largely template wrappers around Evennia defaults.
## Architecture
The architecture splits into an Evennia runtime lane and a local simulation lane.
```mermaid
graph TD
subgraph External Clients
Telnet[Telnet client :4000]
Browser[Browser / webclient :4001]
Operator[Local operator]
end
subgraph Evennia Runtime
Settings[server/conf/settings.py]
URLs[web/urls.py]
Cmdsets[commands/default_cmdsets.py]
Typeclasses[typeclasses/*.py]
WorldDocs[world/prototypes.py + world/help_entries.py]
WebHooks[server/conf/web_plugins.py]
end
subgraph Standalone Tower Simulator
Play200[play_200.py]
RootGame[game.py]
AltGame[world/game.py]
Engine[GameEngine / PlayerInterface / NPCAI]
State[game_state.json + timmy_log.md]
end
Telnet --> Settings
Browser --> URLs
Settings --> Cmdsets
Cmdsets --> Typeclasses
URLs --> WebHooks
Typeclasses --> WorldDocs
Operator --> Play200
Play200 --> RootGame
RootGame --> Engine
AltGame --> Engine
Engine --> State
```
What is actually wired today:
- `server/conf/settings.py` only overrides `SERVERNAME = "timmy_world"` and optionally imports `server.conf.secret_settings`.
- `web/urls.py` mounts `web.website.urls`, `web.webclient.urls`, `web.admin.urls`, then appends `evennia.web.urls`.
- `commands/default_cmdsets.py` subclasses Evennia defaults but does not add custom commands yet.
- `typeclasses/*.py` are thin wrappers around Evennia defaults.
- `server/conf/web_plugins.py` returns the web roots unchanged.
- `server/conf/at_initial_setup.py` is a no-op.
- `world/batch_cmds.ev` is still template commentary rather than a real build script.
What is custom and stateful today:
- `evennia/timmy_world/game.py`
- `evennia/timmy_world/world/game.py`
- `evennia/timmy_world/play_200.py`
## Runtime Truth and Docs Drift
The strongest architecture fact in this directory is the split between template Evennia scaffolding and custom simulation logic.
Drift discovered during inspection:
- `evennia/timmy_world/README.md` is the stock Evennia welcome text.
- `server/conf/at_initial_setup.py` is empty, so the Evennia world is not auto-populating custom Tower content at first boot.
- `world/batch_cmds.ev` is also a template, not a concrete room/object bootstrap file.
- The deepest custom logic is not in the typeclasses or server hooks. It is in `evennia/timmy_world/game.py` and `evennia/timmy_world/world/game.py`.
- `evennia/timmy_world/play_200.py` imports `from game import GameEngine, NARRATIVE_PHASES`, which proves the root `game.py` is an active entry point.
- `evennia/timmy_world/world/game.py` is not dead weight either; it contains its own `World`, `ActionSystem`, `NPCAI`, `DialogueSystem`, `GameEngine`, and `PlayerInterface` stack.
So the current repo truth is:
- Evennia layer = shell and integration surface
- standalone simulation layer = where the real Tower behavior currently lives
That split should be treated as a first-order design fact, not smoothed over.
## Entry Points
### 1. Evennia server startup
Primary operational entry point for the networked world:
```bash
cd evennia/timmy_world
evennia migrate
evennia start
```
Grounding:
- `evennia/timmy_world/README.md`
- `evennia/timmy_world/server/conf/settings.py`
### 2. Web routing
`evennia/timmy_world/web/urls.py` is the browser-facing entry point. It includes:
- `web.website.urls`
- `web.webclient.urls`
- `web.admin.urls`
- `evennia.web.urls` appended after the local patterns
This means the effective surface inherits Evennia defaults rather than defining a custom Tower web application.
### 3. Standalone simulation module
`evennia/timmy_world/game.py` is a pure-Python entry point with:
- `NARRATIVE_PHASES`
- `get_narrative_phase()`
- `get_phase_transition_event()`
- `World`
- `ActionSystem`
- `NPCAI`
- `GameEngine`
- `PlayerInterface`
This module can be imported and exercised without an Evennia runtime.
### 4. Alternate simulation module
`evennia/timmy_world/world/game.py` mirrors much of the same gameplay stack, but is not the one used by `play_200.py`.
Important distinction:
- root `game.py` is the active scripted demo target
- `world/game.py` is a second engine implementation with overlapping responsibilities
### 5. Scripted narrative demo
`evennia/timmy_world/play_200.py` runs 200 deterministic ticks and prints a story arc across four named phases:
- Quietus
- Fracture
- Breaking
- Mending
This file is the clearest executable artifact proving how the simulator is intended to be consumed outside Evennia.
## Data Flow
### Networked Evennia path
1. Client connects via telnet or browser.
2. Evennia loads settings from `server/conf/settings.py`.
3. Command set resolution flows through `commands/default_cmdsets.py`.
4. Typeclass objects resolve through `typeclasses/accounts.py`, `typeclasses/characters.py`, `typeclasses/rooms.py`, `typeclasses/exits.py`, `typeclasses/objects.py`, and `typeclasses/scripts.py`.
5. URL dispatch flows through `web/urls.py` into website, webclient, admin, and Evennia default URL patterns.
6. Object/help/prototype metadata can be sourced from `world/prototypes.py` and `world/help_entries.py`.
### Standalone Tower simulation path
1. Operator imports `evennia/timmy_world/game.py` directly or runs `evennia/timmy_world/play_200.py`.
2. `GameEngine.start_new_game()` initializes the world state.
3. `PlayerInterface.get_available_actions()` exposes current verbs from room topology and nearby characters.
4. `GameEngine.run_tick()` / `play_turn()` advances time, movement, world events, NPC actions, and logs.
5. `World` tracks rooms, characters, trust, weather, forge/garden/bridge/tower state, and narrative phase.
6. Persistence writes to JSON/log files rooted at `/Users/apayne/.timmy/evennia/timmy_world`.
### Evidence of the persistence contract
Both simulation modules hardcode the same portability-sensitive base path:
- `evennia/timmy_world/game.py`
- `evennia/timmy_world/world/game.py`
Each defines:
- `WORLD_DIR = Path('/Users/apayne/.timmy/evennia/timmy_world')`
- `STATE_FILE = WORLD_DIR / 'game_state.json'`
- `TIMMY_LOG = WORLD_DIR / 'timmy_log.md'`
## Key Abstractions
### `World` — state container for the Tower
Found in both `evennia/timmy_world/game.py` and `evennia/timmy_world/world/game.py`.
Responsibilities:
- defines the five-room map: Threshold, Tower, Forge, Garden, Bridge
- stores per-room connections and dynamic state
- stores per-character room, energy, trust, goals, memories, and inventory
- tracks global pressure variables like `forge_fire_dying`, `garden_drought`, `bridge_flooding`, and `tower_power_low`
- updates world time and environmental drift each tick
### `ActionSystem`
Also present in both engine files.
Responsibilities:
- enumerates available verbs
- computes contextual action menus from world state
- ties actions to energy cost and room/character context
### `NPCAI`
The non-player decision layer.
Responsibilities:
- chooses actions based on each character's goals and situation
- creates world motion without requiring live operator input
- in `world/game.py`, works alongside `DialogueSystem`
### `GameEngine`
The orchestration layer.
Responsibilities:
- bootstraps a fresh run with `start_new_game()`
- rehydrates from storage via `load_game()`
- advances the simulation with `run_tick()` / `play_turn()`
- records log entries and world events
Grounded interface details from live import of `evennia/timmy_world/game.py`:
- methods visible on the instance: `load_game`, `log`, `play_turn`, `run_tick`, `start_new_game`
- `play_turn('look')` returns a dict with keys:
- `tick`
- `time`
- `phase`
- `phase_name`
- `timmy_room`
- `timmy_energy`
- `room_desc`
- `here`
- `world_events`
- `npc_actions`
- `choices`
- `log`
### `PlayerInterface`
A thin operator-facing adapter.
Grounded behavior:
- when loaded from `evennia/timmy_world/game.py` after `start_new_game()`, `PlayerInterface(engine).get_available_actions()` exposes room navigation and social verbs like:
- `move:north -> Tower`
- `move:east -> Garden`
- `move:west -> Forge`
- `move:south -> Bridge`
- `speak:Allegro`
- `speak:Claude`
- `rest`
### Evennia typeclasses and cmdsets
The Evennia abstractions are real but thin.
Notable files:
- `evennia/timmy_world/typeclasses/objects.py`
- `evennia/timmy_world/typeclasses/characters.py`
- `evennia/timmy_world/typeclasses/rooms.py`
- `evennia/timmy_world/typeclasses/exits.py`
- `evennia/timmy_world/typeclasses/accounts.py`
- `evennia/timmy_world/typeclasses/scripts.py`
- `evennia/timmy_world/commands/command.py`
- `evennia/timmy_world/commands/default_cmdsets.py`
Today these mostly wrap Evennia defaults instead of implementing a custom Tower-specific protocol on top.
## API Surface
### Network surfaces
Grounded from `README.md`, `web/urls.py`, and `server/conf/mssp.py`:
- Telnet on port `4000`
- Browser / webclient on `http://localhost:4001`
- admin surface under `/admin/`
- Evennia default URLs appended via `evennia.web.urls`
- Evennia REST/web surface inherits the default `/api/` patterns rather than defining custom project-specific endpoints here
### Operator / script surfaces
- `python3 evennia/timmy_world/play_200.py`
- importable pure-Python engine in `evennia/timmy_world/game.py`
- alternate engine in `evennia/timmy_world/world/game.py`
### Content/model surfaces
- object prototype definitions: `evennia/timmy_world/world/prototypes.py`
- file-based help entries: `evennia/timmy_world/world/help_entries.py`
## Test Coverage Gaps
### Current verified state
The original genome here was stale. The live repo now shows two different categories of test coverage:
1. Host-repo generated tests already exist in `tests/test_genome_generated.py`
- they reference `evennia/timmy_world/game.py`
- they reference `evennia/timmy_world/world/game.py`
- they reference `server/conf/web_plugins.py`
2. Those generated tests are not trustworthy as-is for this target
- running `python3 -m pytest tests/test_genome_generated.py -k 'EvenniaTimmyWorld' -q -rs`
- result: `19 skipped, 31 deselected`
- skip reason on every case: `Module not importable`
This matters because the codebase-genome pipeline reported zero local tests for the subproject, but the host repo does contain tests. The real issue is not “no tests exist.” The real issue is “the existing generated tests are disconnected from the actual import path and therefore do not execute the critical path.”
### New critical-path tests added for #677
This issue refresh adds a dedicated executable test file:
- `tests/test_evennia_local_world_game.py`
Covered behaviors:
- narrative phase boundaries across Quietus / Fracture / Breaking / Mending
- player-facing action surface from the Threshold start state
- deterministic `run_tick('move:north')` flow into the Tower with expected log and world-event output
### Genome artifact coverage added for #677
This issue refresh also adds:
- `tests/test_evennia_local_world_genome.py`
That test locks:
- artifact path
- required analysis sections
- grounded snippets for real files and verification output
### Remaining gaps
Still missing strong runtime coverage for:
- Evennia typeclass behavior under a real Evennia test harness
- URL routing under Django/Evennia integration
- `world/game.py` parity versus root `game.py`
- persistence portability around `/Users/apayne/.timmy/evennia/timmy_world`
- `at_initial_setup.py` and `world/batch_cmds.ev` actually building a playable world in the Evennia path
## Security Considerations
1. Plaintext telnet exposure
- `server/conf/mssp.py` advertises port `4000`
- telnet is unencrypted by default
- acceptable for localhost/dev, risky for exposed deployment
2. Hardcoded absolute persistence path
- both `evennia/timmy_world/game.py` and `evennia/timmy_world/world/game.py` hardcode `/Users/apayne/.timmy/evennia/timmy_world`
- this couples runtime writes to one operator machine and one home-directory layout
- portability and accidental overwrite risk are both real
- filed follow-up: `timmy-home #831``https://forge.alexanderwhitestone.com/Timmy_Foundation/timmy-home/issues/831`
3. Admin/web surfaces inherit defaults
- `web/urls.py` exposes admin and Evennia defaults
- if the service is made remotely reachable, Django/Evennia auth and proxy boundaries matter immediately
4. Secret handling is externalized but optional
- `server/conf/settings.py` silently falls back if `secret_settings.py` is missing
- convenient for local development, but secrets discipline lives outside the repo contract
5. Template hooks can hide missing security posture
- `server/conf/web_plugins.py` is pass-through
- `server/conf/at_initial_setup.py` is pass-through
- the absence of custom code here means there are no local hardening hooks yet for startup, proxying, or world bootstrap
## Dependencies
Directly evidenced imports and framework coupling:
- Evennia 6.0 game-directory structure
- Django via Evennia web/admin stack
- Twisted via Evennia networking/web hooks
- Python stdlib heavy use in standalone simulator:
- `json`
- `time`
- `os`
- `random`
- `datetime`
- `pathlib`
- `sys`
Dependency caveat:
- the standalone Tower simulator is largely pure Python and importable in isolation
- the typeclass / cmdset / web files depend on Evennia and Django runtime wiring to do real work
## Deployment
### Evennia path
```bash
cd evennia/timmy_world
evennia migrate
evennia start
```
Expected local surfaces from repo docs/config:
- telnet: `localhost:4000`
- browser/webclient: `http://localhost:4001`
### Standalone simulation path
```bash
cd evennia/timmy_world
python3 play_200.py
```
This does not require the full Evennia network stack. It exercises the root `game.py` engine directly.
### Verification commands run for this genome refresh
```bash
python3 ~/.hermes/pipelines/codebase-genome.py --path /tmp/BURN-7-7/evennia/timmy_world --output /tmp/evennia-local-world-GENOME-base.md
python3 -m pytest tests/test_genome_generated.py -k 'EvenniaTimmyWorld' -q -rs
python3 -m pytest tests/test_evennia_local_world_genome.py tests/test_evennia_local_world_game.py -q
python3 -m py_compile evennia/timmy_world/game.py evennia/timmy_world/world/game.py evennia/timmy_world/play_200.py evennia/timmy_world/server/conf/settings.py evennia/timmy_world/web/urls.py
```
## Key Findings
1. The current custom product logic is the standalone Tower simulator, not the Evennia typeclass layer.
2. The repo contains two parallel simulation engines: `evennia/timmy_world/game.py` and `evennia/timmy_world/world/game.py`.
3. The stock Evennia scaffolding is still mostly template code (`README.md`, `at_initial_setup.py`, `world/batch_cmds.ev`, pass-through cmdsets/web hooks).
4. The codebase-genome pipeline undercounted test reality because subproject-local tests are absent while host-repo tests exist one level up.
5. The existing generated tests were present but functionally inert: `19 skipped` because their import path does not match the current host-repo layout.
6. The most concrete portability hazard is the hardcoded `/Users/apayne/.timmy/evennia/timmy_world` state path in both simulation engines.
---
This refreshed genome supersedes the earlier auto-generated `evennia/timmy_world/GENOME.md` summary by grounding the analysis in live source inspection, live import of `evennia/timmy_world/game.py`, current file metrics, and executable host-repo verification.

View File

@@ -0,0 +1,110 @@
#
# Bezalel World Builder — Evennia batch commands
# Creates the Bezalel Evennia world from evennia_tools/bezalel_layout.py specs.
#
# Load with: @batchcommand bezalel_world
#
# Part of #536
# Create rooms
@create/drop Limbo:evennia.objects.objects.DefaultRoom
@desc here = The void between worlds. The air carries the pulse of three houses: Mac, VPS, and this one. Everything begins here before it is given form.
@create/drop Gatehouse:evennia.objects.objects.DefaultRoom
@desc here = A stone guard tower at the edge of Bezalel world. The walls are carved with runes of travel, proof, and return. Every arrival is weighed before it is trusted.
@create/drop Great Hall:evennia.objects.objects.DefaultRoom
@desc here = A vast hall with a long working table. Maps of the three houses hang beside sketches, benchmarks, and deployment notes. This is where the forge reports back to the house.
@create/drop The Library of Bezalel:evennia.objects.objects.DefaultRoom
@desc here = Shelves of technical manuals, Evennia code, test logs, and bridge schematics rise to the ceiling. This room holds plans waiting to be made real.
@create/drop The Observatory:evennia.objects.objects.DefaultRoom
@desc here = A high chamber with telescopes pointing toward the Mac, the VPS, and the wider net. Screens glow with status lights, latency traces, and long-range signals.
@create/drop The Workshop:evennia.objects.objects.DefaultRoom
@desc here = A forge and workbench share the same heat. Scattered here are half-finished bridges, patched harnesses, and tools laid out for proof before pride.
@create/drop The Server Room:evennia.objects.objects.DefaultRoom
@desc here = Racks of humming servers line the walls. Fans push warm air through the chamber while status LEDs beat like a mechanical heart. This is the pulse of Bezalel house.
@create/drop The Garden of Code:evennia.objects.objects.DefaultRoom
@desc here = A quiet garden where ideas are left long enough to grow roots. Code-shaped leaves flutter in patterned wind, and a stone path invites patient thought.
@create/drop The Portal Room:evennia.objects.objects.DefaultRoom
@desc here = Three shimmering doorways stand in a ring: one marked for the Mac house, one for the VPS, and one for the wider net. The room hums like a bridge waiting for traffic.
# Create exits
@open gatehouse:gate,tower = Gatehouse
@open limbo:void,back = Limbo
@open greathall:hall,great hall = Great Hall
@open gatehouse:gate,tower = Gatehouse
@open library:books,study = The Library of Bezalel
@open hall:great hall,back = Great Hall
@open observatory:telescope,tower top = The Observatory
@open hall:great hall,back = Great Hall
@open workshop:forge,bench = The Workshop
@open hall:great hall,back = Great Hall
@open serverroom:servers,server room = The Server Room
@open workshop:forge,bench = The Workshop
@open garden:garden of code,grove = The Garden of Code
@open workshop:forge,bench = The Workshop
@open portalroom:portal,portals = The Portal Room
@open gatehouse:gate,back = Gatehouse
# Create objects
@create Threshold Ledger
@desc Threshold Ledger = A heavy ledger where arrivals, departures, and field notes are recorded before the work begins.
@tel Threshold Ledger = Gatehouse
@create Three-House Map
@desc Three-House Map = A long map showing Mac, VPS, and remote edges in one continuous line of work.
@tel Three-House Map = Great Hall
@create Bridge Schematics
@desc Bridge Schematics = Rolled plans describing world bridges, Evennia layouts, and deployment paths.
@tel Bridge Schematics = The Library of Bezalel
@create Compiler Manuals
@desc Compiler Manuals = Manuals annotated in the margins with warnings against cleverness without proof.
@tel Compiler Manuals = The Library of Bezalel
@create Tri-Axis Telescope
@desc Tri-Axis Telescope = A brass telescope assembly that can be turned toward the Mac, the VPS, or the open net.
@tel Tri-Axis Telescope = The Observatory
@create Forge Anvil
@desc Forge Anvil = Scarred metal used for turning rough plans into testable form.
@tel Forge Anvil = The Workshop
@create Bridge Workbench
@desc Bridge Workbench = A wide bench covered in harness patches, relay notes, and half-soldered bridge parts.
@tel Bridge Workbench = The Workshop
@create Heartbeat Console
@desc Heartbeat Console = A monitoring console showing service health, latency, and the steady hum of the house.
@tel Heartbeat Console = The Server Room
@create Server Racks
@desc Server Racks = Stacked machines that keep the world awake even when no one is watching.
@tel Server Racks = The Server Room
@create Code Orchard
@desc Code Orchard = Trees with code-shaped leaves. Some branches bear elegant abstractions; others hold broken prototypes.
@tel Code Orchard = The Garden of Code
@create Stone Bench
@desc Stone Bench = A place to sit long enough for a hard implementation problem to become clear.
@tel Stone Bench = The Garden of Code
@create Mac Portal:mac arch
@desc Mac Portal = A silver doorway whose frame vibrates with the local sovereign house.
@tel Mac Portal = The Portal Room
@create VPS Portal:vps arch
@desc VPS Portal = A cobalt doorway tuned toward the testbed VPS house.
@tel VPS Portal = The Portal Room
@create Net Portal:net arch,network arch
@desc Net Portal = A pale doorway pointed toward the wider net and every uncertain edge beyond it.
@tel Net Portal = The Portal Room

View File

@@ -0,0 +1,190 @@
from collections import deque
from dataclasses import dataclass
@dataclass(frozen=True)
class RoomSpec:
key: str
desc: str
@dataclass(frozen=True)
class ExitSpec:
source: str
key: str
destination: str
aliases: tuple[str, ...] = ()
@dataclass(frozen=True)
class ObjectSpec:
key: str
location: str
desc: str
aliases: tuple[str, ...] = ()
@dataclass(frozen=True)
class CharacterSpec:
key: str
desc: str
starting_room: str
@dataclass(frozen=True)
class TravelCommandSpec:
key: str
aliases: tuple[str, ...]
target_world: str
fallback_room: str
desc: str
ROOMS = (
RoomSpec(
"Limbo",
"The void between worlds. The air carries the pulse of three houses: Mac, VPS, and this one. "
"Everything begins here before it is given form.",
),
RoomSpec(
"Gatehouse",
"A stone guard tower at the edge of Bezalel's world. The walls are carved with runes of travel, "
"proof, and return. Every arrival is weighed before it is trusted.",
),
RoomSpec(
"Great Hall",
"A vast hall with a long working table. Maps of the three houses hang beside sketches, benchmarks, "
"and deployment notes. This is where the forge reports back to the house.",
),
RoomSpec(
"The Library of Bezalel",
"Shelves of technical manuals, Evennia code, test logs, and bridge schematics rise to the ceiling. "
"This room holds plans waiting to be made real.",
),
RoomSpec(
"The Observatory",
"A high chamber with telescopes pointing toward the Mac, the VPS, and the wider net. Screens glow with "
"status lights, latency traces, and long-range signals.",
),
RoomSpec(
"The Workshop",
"A forge and workbench share the same heat. Scattered here are half-finished bridges, patched harnesses, "
"and tools laid out for proof before pride.",
),
RoomSpec(
"The Server Room",
"Racks of humming servers line the walls. Fans push warm air through the chamber while status LEDs beat "
"like a mechanical heart. This is the pulse of Bezalel's house.",
),
RoomSpec(
"The Garden of Code",
"A quiet garden where ideas are left long enough to grow roots. Code-shaped leaves flutter in patterned wind, "
"and a stone path invites patient thought.",
),
RoomSpec(
"The Portal Room",
"Three shimmering doorways stand in a ring: one marked for the Mac house, one for the VPS, and one for the wider net. "
"The room hums like a bridge waiting for traffic.",
),
)
EXITS = (
ExitSpec("Limbo", "gatehouse", "Gatehouse", ("gate", "tower")),
ExitSpec("Gatehouse", "limbo", "Limbo", ("void", "back")),
ExitSpec("Gatehouse", "greathall", "Great Hall", ("hall", "great hall")),
ExitSpec("Great Hall", "gatehouse", "Gatehouse", ("gate", "tower")),
ExitSpec("Great Hall", "library", "The Library of Bezalel", ("books", "study")),
ExitSpec("The Library of Bezalel", "hall", "Great Hall", ("great hall", "back")),
ExitSpec("Great Hall", "observatory", "The Observatory", ("telescope", "tower top")),
ExitSpec("The Observatory", "hall", "Great Hall", ("great hall", "back")),
ExitSpec("Great Hall", "workshop", "The Workshop", ("forge", "bench")),
ExitSpec("The Workshop", "hall", "Great Hall", ("great hall", "back")),
ExitSpec("The Workshop", "serverroom", "The Server Room", ("servers", "server room")),
ExitSpec("The Server Room", "workshop", "The Workshop", ("forge", "bench")),
ExitSpec("The Workshop", "garden", "The Garden of Code", ("garden of code", "grove")),
ExitSpec("The Garden of Code", "workshop", "The Workshop", ("forge", "bench")),
ExitSpec("Gatehouse", "portalroom", "The Portal Room", ("portal", "portals")),
ExitSpec("The Portal Room", "gatehouse", "Gatehouse", ("gate", "back")),
)
OBJECTS = (
ObjectSpec("Threshold Ledger", "Gatehouse", "A heavy ledger where arrivals, departures, and field notes are recorded before the work begins."),
ObjectSpec("Three-House Map", "Great Hall", "A long map showing Mac, VPS, and remote edges in one continuous line of work."),
ObjectSpec("Bridge Schematics", "The Library of Bezalel", "Rolled plans describing world bridges, Evennia layouts, and deployment paths."),
ObjectSpec("Compiler Manuals", "The Library of Bezalel", "Manuals annotated in the margins with warnings against cleverness without proof."),
ObjectSpec("Tri-Axis Telescope", "The Observatory", "A brass telescope assembly that can be turned toward the Mac, the VPS, or the open net."),
ObjectSpec("Forge Anvil", "The Workshop", "Scarred metal used for turning rough plans into testable form."),
ObjectSpec("Bridge Workbench", "The Workshop", "A wide bench covered in harness patches, relay notes, and half-soldered bridge parts."),
ObjectSpec("Heartbeat Console", "The Server Room", "A monitoring console showing service health, latency, and the steady hum of the house."),
ObjectSpec("Server Racks", "The Server Room", "Stacked machines that keep the world awake even when no one is watching."),
ObjectSpec("Code Orchard", "The Garden of Code", "Trees with code-shaped leaves. Some branches bear elegant abstractions; others hold broken prototypes."),
ObjectSpec("Stone Bench", "The Garden of Code", "A place to sit long enough for a hard implementation problem to become clear."),
ObjectSpec("Mac Portal", "The Portal Room", "A silver doorway whose frame vibrates with the local sovereign house.", ("mac arch",)),
ObjectSpec("VPS Portal", "The Portal Room", "A cobalt doorway tuned toward the testbed VPS house.", ("vps arch",)),
ObjectSpec("Net Portal", "The Portal Room", "A pale doorway pointed toward the wider net and every uncertain edge beyond it.", ("net arch", "network arch")),
)
CHARACTERS = (
CharacterSpec("Timmy", "The Builder's first creation. Quiet, observant, already measuring the room before he speaks.", "Gatehouse"),
CharacterSpec("Bezalel", "The forge-and-testbed wizard. Scarred hands, steady gaze, the habit of proving things before trusting them.", "The Workshop"),
CharacterSpec("Marcus", "An old man with kind eyes. He walks like someone who has already survived the night once.", "The Garden of Code"),
CharacterSpec("Kimi", "The deep scholar of context and meaning. He carries long memory like a lamp.", "The Library of Bezalel"),
)
PORTAL_COMMANDS = (
TravelCommandSpec(
"mac",
("macbook", "local"),
"Mac house",
"Limbo",
"Align with the sovereign local house. Until live cross-world transport is wired, the command resolves into Limbo — the threshold between houses.",
),
TravelCommandSpec(
"vps",
("testbed", "house"),
"VPS house",
"Limbo",
"Step toward the forge VPS. For now the command lands in Limbo, preserving the inter-world threshold until real linking is live.",
),
TravelCommandSpec(
"net",
("network", "wider-net"),
"Wider net",
"Limbo",
"Face the open network. The command currently routes through Limbo so the direction exists before the final bridge does.",
),
)
def room_keys() -> tuple[str, ...]:
return tuple(room.key for room in ROOMS)
def character_keys() -> tuple[str, ...]:
return tuple(character.key for character in CHARACTERS)
def portal_command_keys() -> tuple[str, ...]:
return tuple(command.key for command in PORTAL_COMMANDS)
def grouped_exits() -> dict[str, tuple[ExitSpec, ...]]:
grouped: dict[str, list[ExitSpec]] = {}
for exit_spec in EXITS:
grouped.setdefault(exit_spec.source, []).append(exit_spec)
return {key: tuple(value) for key, value in grouped.items()}
def reachable_rooms_from(start: str) -> set[str]:
seen: set[str] = set()
queue: deque[str] = deque([start])
exits_by_room = grouped_exits()
while queue:
current = queue.popleft()
if current in seen:
continue
seen.add(current)
for exit_spec in exits_by_room.get(current, ()):
if exit_spec.destination not in seen:
queue.append(exit_spec.destination)
return seen

View File

@@ -0,0 +1,85 @@
#!/usr/bin/env python3
""
build_bezalel_world.py Build Bezalel Evennia world from layout specs.
Programmatically creates rooms, exits, objects, and characters in a running
Evennia instance using the specs from evennia_tools/bezalel_layout.py.
Usage (in Evennia game shell):
from evennia_tools.build_bezalel_world import build_world
build_world()
Or via batch command:
@batchcommand evennia_tools/batch_cmds_bezalel.ev
Part of #536
""
from evennia_tools.bezalel_layout import (
ROOMS, EXITS, OBJECTS, CHARACTERS, PORTAL_COMMANDS,
room_keys, reachable_rooms_from
)
def build_world():
"""Build the Bezalel Evennia world from layout specs."""
from evennia.objects.models import ObjectDB
from evennia.utils.create import create_object, create_exit, create_message
print("Building Bezalel world...")
# Create rooms
rooms = {}
for spec in ROOMS:
room = create_object(
"evennia.objects.objects.DefaultRoom",
key=spec.key,
attributes=(("desc", spec.desc),),
)
rooms[spec.key] = room
print(f" Room: {spec.key}")
# Create exits
for spec in EXITS:
source = rooms.get(spec.source)
dest = rooms.get(spec.destination)
if not source or not dest:
print(f" WARNING: Exit {spec.key} — missing room")
continue
exit_obj = create_exit(
key=spec.key,
location=source,
destination=dest,
aliases=list(spec.aliases),
)
print(f" Exit: {spec.source} -> {spec.destination} ({spec.key})")
# Create objects
for spec in OBJECTS:
location = rooms.get(spec.location)
if not location:
print(f" WARNING: Object {spec.key} — missing room {spec.location}")
continue
obj = create_object(
"evennia.objects.objects.DefaultObject",
key=spec.key,
location=location,
attributes=(("desc", spec.desc),),
aliases=list(spec.aliases),
)
print(f" Object: {spec.key} in {spec.location}")
# Verify reachability
all_rooms = set(room_keys())
reachable = reachable_rooms_from("Limbo")
unreachable = all_rooms - reachable
if unreachable:
print(f" WARNING: Unreachable rooms: {unreachable}")
else:
print(f" All {len(all_rooms)} rooms reachable from Limbo")
print("Bezalel world built.")
if __name__ == "__main__":
build_world()

View File

@@ -0,0 +1,270 @@
from __future__ import annotations
from dataclasses import asdict, dataclass
HALL_OF_KNOWLEDGE = "Hall of Knowledge"
LEDGER_OBJECT = "The Ledger"
@dataclass(frozen=True)
class MindPalaceIssue:
issue_number: int
state: str
title: str
layer: str
spatial_role: str
rationale: str
def summary_line(self) -> str:
return f"#{self.issue_number} {self.title} [{self.state} · {self.layer} · {self.spatial_role}]"
@dataclass(frozen=True)
class MutableFact:
key: str
value: str
source: str
def to_dict(self) -> dict[str, str]:
return asdict(self)
@dataclass(frozen=True)
class BurnCycleSnapshot:
repo: str
branch: str
active_issue: int
focus: str
active_operator: str
blockers: tuple[str, ...] = ()
def to_dict(self) -> dict[str, object]:
return {
"repo": self.repo,
"branch": self.branch,
"active_issue": self.active_issue,
"focus": self.focus,
"active_operator": self.active_operator,
"blockers": list(self.blockers),
}
EVENNIA_MIND_PALACE_ISSUES = (
MindPalaceIssue(
508,
"closed",
"[P0] Tower Game — contextual dialogue (NPCs recycle 15 lines forever)",
"L4",
"Dialogue tutor NPCs",
"Contextual dialogue belongs in procedural behavior surfaces so the right NPC can teach or respond based on current room state.",
),
MindPalaceIssue(
509,
"closed",
"[P0] Tower Game — trust must decrease, conflict must exist",
"L2",
"Mutable relationship state",
"Trust, resentment, and alliance changes are world facts that should live on objects and characters, not in flat prompt text.",
),
MindPalaceIssue(
510,
"closed",
"[P0] Tower Game — narrative arc (tick 200 = tick 20)",
"L3",
"Archive chronicle",
"A spatial memory needs a chronicle room where prior events can be replayed and searched so the world can develop an actual arc.",
),
MindPalaceIssue(
511,
"open",
"[P0] Tower Game — energy must meaningfully constrain",
"L2",
"Mutable world meter",
"Energy is a changing state variable that should be visible in-room and affect what actions remain possible.",
),
MindPalaceIssue(
512,
"open",
"[P1] Sonnet workforce — full end-to-end smoke test",
"L3",
"Proof shelf",
"End-to-end smoke traces belong in the archive so world behavior can be proven, revisited, and compared over time.",
),
MindPalaceIssue(
513,
"open",
"[P1] Tower Game — world events must affect gameplay",
"L2",
"Event-reactive room state",
"If storms, fire, or decay do not alter the room state, the world is decorative instead of mnemonic.",
),
MindPalaceIssue(
514,
"open",
"[P1] Tower Game — items that change the world",
"L2",
"Interactive objects",
"World-changing items are exactly the kind of mutable facts and affordances that a spatial memory substrate should expose.",
),
MindPalaceIssue(
515,
"open",
"[P1] Tower Game — NPC-NPC relationships",
"L2",
"Social graph in-world",
"Relationships should persist on characters and become inspectable through spatial proximity rather than hidden transcript-only state.",
),
MindPalaceIssue(
516,
"closed",
"[P1] Tower Game — Timmy richer dialogue + internal monologue",
"L4",
"Inner-room teaching patterns",
"Internal monologue and richer dialogue are procedural behaviors that can be attached to rooms, NPCs, and character routines.",
),
MindPalaceIssue(
517,
"open",
"[P1] Tower Game — NPCs move between rooms with purpose",
"L5",
"Movement-driven retrieval",
"Purposeful movement is retrieval logic made spatial: who enters which room determines what knowledge is loaded and acted on.",
),
MindPalaceIssue(
534,
"open",
"[BEZ-P0] Fix Evennia settings on 104.131.15.18 — remove bad port tuples, DB is ready",
"L1",
"Runtime threshold",
"Before the mind palace can be inhabited, the base Evennia runtime topology has to load cleanly at the threshold.",
),
MindPalaceIssue(
535,
"open",
"[BEZ-P0] Install Tailscale on Bezalel VPS (104.131.15.18) for internal networking",
"L1",
"Network threshold",
"Network identity and reachability are static environment facts that determine which rooms and worlds are even reachable.",
),
MindPalaceIssue(
536,
"open",
"[BEZ-P1] Create Bezalel Evennia world with themed rooms and characters",
"L1",
"First room graph",
"Themed rooms and characters are the static world scaffold that lets memory become place instead of prose.",
),
MindPalaceIssue(
537,
"closed",
"[BRIDGE-P1] Deploy Evennia bridge API on all worlds — sync presence and events",
"L5",
"Cross-world routing",
"Bridge APIs turn movement across worlds into retrieval across houses instead of forcing one global prompt blob.",
),
MindPalaceIssue(
538,
"closed",
"[ALLEGRO-P1] Fix SSH access from Mac to Allegro VPS (167.99.126.228)",
"L1",
"Operator ingress",
"Operator access is part of the static world boundary: if the house cannot be reached, its memory cannot be visited.",
),
MindPalaceIssue(
539,
"closed",
"[ARCH-P2] Implement Evennia hub-and-spoke federation architecture",
"L5",
"Federated retrieval map",
"Federation turns room-to-room travel into selective retrieval across sovereign worlds instead of a single central cache.",
),
)
OPEN_EVENNIA_MIND_PALACE_ISSUES = tuple(issue for issue in EVENNIA_MIND_PALACE_ISSUES if issue.state == "open")
def build_hall_of_knowledge_entry(
active_issues: tuple[MindPalaceIssue, ...] | list[MindPalaceIssue],
ledger_fact: MutableFact,
burn_cycle: BurnCycleSnapshot,
) -> dict[str, object]:
issue_lines = [issue.summary_line() for issue in active_issues]
blocker_lines = list(burn_cycle.blockers) or ["No blockers recorded."]
return {
"room": {
"key": HALL_OF_KNOWLEDGE,
"purpose": "Load live issue topology, current burn-cycle focus, and the minimum durable facts Timmy needs before acting.",
},
"object": {
"key": LEDGER_OBJECT,
"purpose": "Expose one mutable fact from Timmy's durable memory so the room proves stateful recall instead of static documentation.",
"fact": ledger_fact.to_dict(),
},
"ambient_context": [
f"Room entry into {HALL_OF_KNOWLEDGE} preloads active Gitea issue topology for {burn_cycle.repo}.",
*issue_lines,
f"Ledger fact {ledger_fact.key}: {ledger_fact.value}",
f"Timmy burn cycle focus: issue #{burn_cycle.active_issue} on {burn_cycle.branch}{burn_cycle.focus}",
f"Operator lane: {burn_cycle.active_operator}",
],
"burn_cycle": burn_cycle.to_dict(),
"commands": {
"/who lives here": "; ".join(issue_lines) or "No issues loaded.",
"/status forge": f"{burn_cycle.repo} @ {burn_cycle.branch} (issue #{burn_cycle.active_issue})",
"/what is broken": "; ".join(blocker_lines),
},
}
def render_room_entry_proof(
active_issues: tuple[MindPalaceIssue, ...] | list[MindPalaceIssue],
ledger_fact: MutableFact,
burn_cycle: BurnCycleSnapshot,
) -> str:
entry = build_hall_of_knowledge_entry(active_issues, ledger_fact, burn_cycle)
lines = [
f"ENTER {entry['room']['key']}",
f"Purpose: {entry['room']['purpose']}",
"Ambient context:",
]
lines.extend(f"- {line}" for line in entry["ambient_context"])
lines.extend(
[
f"Object: {entry['object']['key']}",
f"- {entry['object']['fact']['key']}: {entry['object']['fact']['value']}",
f"- source: {entry['object']['fact']['source']}",
"Timmy burn cycle:",
f"- repo: {burn_cycle.repo}",
f"- branch: {burn_cycle.branch}",
f"- active issue: #{burn_cycle.active_issue}",
f"- focus: {burn_cycle.focus}",
f"- operator: {burn_cycle.active_operator}",
"Command surfaces:",
f"- /who lives here -> {entry['commands']['/who lives here']}",
f"- /status forge -> {entry['commands']['/status forge']}",
f"- /what is broken -> {entry['commands']['/what is broken']}",
]
)
return "\n".join(lines)
def demo_room_entry_proof() -> str:
return render_room_entry_proof(
active_issues=OPEN_EVENNIA_MIND_PALACE_ISSUES[:3],
ledger_fact=MutableFact(
key="canonical-evennia-body",
value="timmy_world on localhost:4001 remains the canonical local body while room entry preloads live issue topology.",
source="reports/production/2026-03-28-evennia-world-proof.md",
),
burn_cycle=BurnCycleSnapshot(
repo="Timmy_Foundation/timmy-home",
branch="fix/567",
active_issue=567,
focus="Evennia as Agent Mind Palace — Spatial Memory Architecture",
active_operator="BURN-7-1",
blockers=("Comment on issue #567 with room-entry proof after PR creation",),
),
)

View File

@@ -0,0 +1,476 @@
# GENOME.md: burn-fleet
**Generated:** 2026-04-15
**Repo:** Timmy_Foundation/burn-fleet
**Purpose:** Laned tmux dispatcher for sovereign burn operations across Mac and Allegro
**Analyzed commit:** `2d4d9ab`
**Size:** 5 top-level source/config files + README | 985 total lines (`fleet-dispatch.py` 320, `fleet-christen.py` 205, `fleet-status.py` 143, `fleet-launch.sh` 126, `fleet-spec.json` 98, `README.md` 93)
---
## Project Overview
`burn-fleet` is a compact control-plane repo for the Hundred-Pane Fleet.
Its job is not model inference itself. Its job is to shape where inference runs, which panes wake up, which repos route to which windows, and how work is fanned out across Mac and VPS workers.
The repo turns a narrative naming scheme into executable infrastructure:
- Mac runs the local session (`BURN`) with windows like `CRUCIBLE`, `GNOMES`, `LOOM`, `FOUNDRY`, `WARD`, `COUNCIL`
- Allegro runs a remote session (`BURN`) with windows like `FORGE`, `ANVIL`, `CRUCIBLE-2`, `SENTINEL`
- `fleet-spec.json` is the single source of truth for pane counts, lanes, sublanes, glyphs, and names
- `fleet-launch.sh` materializes the tmux topology
- `fleet-christen.py` boots `hermes chat --yolo` in each pane and pushes identity prompts
- `fleet-dispatch.py` consumes Gitea issues, maps repos to windows through `MAC_ROUTE` and `ALLEGRO_ROUTE`, and sends `/queue` work into the right panes
- `fleet-status.py` inspects pane output and reports fleet health
The repo is small, but it sits on a high-blast-radius operational seam:
- it controls 100+ panes
- it writes to live tmux sessions
- it comments on live Gitea issues
- it depends on SSH reachability to the VPS
- it is effectively a narrative infrastructure orchestrator
This means the right way to read it is as a dispatch kernel, not just a set of scripts.
---
## Architecture
```mermaid
graph TD
A[fleet-spec.json] --> B[fleet-launch.sh]
A --> C[fleet-christen.py]
A --> D[fleet-dispatch.py]
A --> E[fleet-status.py]
B --> F[tmux session BURN on Mac]
B --> G[tmux session BURN on Allegro over SSH]
C --> F
C --> G
C --> H[hermes chat --yolo in every pane]
H --> I[identity + lane prompt]
J[Gitea issues on forge.alexanderwhitestone.com] --> D
D --> K[MAC_ROUTE]
D --> L[ALLEGRO_ROUTE]
D --> M[/queue prompt generation]
M --> F
M --> G
D --> N[comment_on_issue]
N --> J
D --> O[dispatch-state.json]
E --> F
E --> G
E --> P[get_pane_status]
P --> Q[fleet health summary]
```
### Structural reading
The repo has one real architecture pattern:
1. declarative topology in `fleet-spec.json`
2. imperative realization scripts that consume that topology
3. runtime state in `dispatch-state.json`
4. external side effects in tmux, SSH, and Gitea
That makes `fleet-spec.json` the nucleus and the four scripts adapters around it.
---
## Entry Points
| Entry point | Type | Role |
|-------------|------|------|
| `fleet-launch.sh [mac|allegro|both]` | Shell CLI | Creates tmux sessions and pane layouts from `fleet-spec.json` |
| `python3 fleet-christen.py [mac|allegro|both]` | Python CLI | Starts Hermes workers and injects identity/lane prompts |
| `python3 fleet-dispatch.py [--cycles N] [--interval S] [--machine mac|allegro|both]` | Python CLI | Pulls open Gitea issues, routes them, comments on issues, persists `dispatch-state.json` |
| `python3 fleet-status.py [--machine mac|allegro|both]` | Python CLI | Samples pane output and reports working/idle/error/dead state |
| `README.md` quick start | Human runbook | Documents the intended operator flow from launch to christening to dispatch to status |
### Hidden operational entry points
These are not CLI entry points, but they matter for behavior:
- `MAC_ROUTE` in `fleet-dispatch.py`
- `ALLEGRO_ROUTE` in `fleet-dispatch.py`
- `SKIP_LABELS` and `INACTIVE` filtering in `fleet-dispatch.py`
- `send_to_pane()` as the effectful dispatch primitive
- `comment_on_issue()` as the visible acknowledgement primitive
- `get_pane_status()` in `fleet-status.py` as the fleet health classifier
---
## Data Flow
### 1. Topology creation
`fleet-launch.sh` reads `fleet-spec.json`, parses each window's pane count, and creates the tmux layout.
Flow:
- load spec file path from `SCRIPT_DIR/fleet-spec.json`
- parse `machines.mac.windows` or `machines.allegro.windows`
- create `BURN` session locally or remotely
- create first window, then split panes, then create remaining windows
- continuously tile after splits
This script is layout-only. It does not launch Hermes.
### 2. Agent wake-up / identity seeding
`fleet-christen.py` reads the same `fleet-spec.json` and sends `hermes chat --yolo` into each pane.
After a fixed wait window, it sends a second `/queue` identity message containing:
- glyph
- pane name
- machine name
- window name
- pane number
- sublane
- sovereign operating instructions
That identity message is the bridge from infrastructure to narrative.
The worker is not just launched; it is assigned a mythic/operator identity with a lane.
### 3. Issue harvest and lane dispatch
`fleet-dispatch.py` is the center of the runtime.
Flow:
- load `fleet-spec.json`
- load `dispatch-state.json`
- load Gitea token
- fetch open issues per repo with `requests`
- filter PRs, meta labels, and previously dispatched issues
- build a candidate pool per machine/window
- assign issues pane-by-pane
- call `send_to_pane()` to inject `/queue ...`
- call `comment_on_issue()` to leave a visible burn dispatch comment
- persist the issue assignment into `dispatch-state.json`
Important: the data flow is not issue -> worker directly.
It is:
issue -> repo route table -> window -> pane -> `/queue` prompt -> worker.
### 4. Health sampling
`fleet-status.py` runs the inverse direction.
It samples pane output through `tmux capture-pane` locally or over SSH and classifies the last visible signal as:
- `working`
- `idle`
- `error`
- `dead`
It then summarizes by window, machine, and global fleet totals.
### 5. Runtime state persistence
`dispatch-state.json` is not checked in, but it is the only persistent memory of what the dispatcher already assigned.
That means the runtime depends on a local mutable file rather than a centralized dispatch ledger.
---
## Key Abstractions
### 1. `fleet-spec.json`
This is the primary abstraction in the repo.
It encodes:
- machine identity (`mac`, `allegro`)
- host / SSH details
- hardware metadata (`cores`, `ram_gb`)
- tmux session names
- default model/provider metadata
- windows with `panes`, `lane`, `sublanes`, `glyphs`, `names`
Everything else in the repo interprets this document.
If the spec drifts from the route tables or runtime assumptions, the fleet silently degrades.
### 2. Route tables: `MAC_ROUTE` and `ALLEGRO_ROUTE`
These tables are the repo's second control nucleus.
They map repo names to windows.
This is how `timmy-home`, `the-nexus`, `the-door`, `fleet-ops`, and `the-beacon` land in different operational lanes.
This split means routing logic is duplicated:
- once in the topology spec
- once in Python route dictionaries
That duplication is one of the most important maintainability risks in the repo.
### 3. Pane effect primitive: `send_to_pane()`
`send_to_pane()` is the real actuator.
It turns a dispatch decision into a tmux `send-keys` side effect.
It handles both:
- local tmux injection
- remote SSH + tmux injection
Everything operationally dangerous funnels through this function.
It is therefore a critical path even though the repo has no tests around it.
### 4. Issue acknowledgement primitive: `comment_on_issue()`
This is the repo's social trace primitive.
It posts a burn dispatch comment back to the issue so humans can see that the fleet claimed it.
This is the visible heartbeat of autonomous dispatch.
### 5. Runtime memory: `dispatch-state.json`
This file is the anti-duplication ledger for dispatch cycles.
Without it, the dispatcher would keep recycling the same issues every pass.
Because it is local-file state instead of centralized state, machine locality matters.
### 6. Health classifier: `get_pane_status()`
`fleet-status.py` does not know the true worker state.
It infers state from captured pane output using string heuristics.
So `get_pane_status()` is effectively a lightweight log classifier.
Its correctness depends on fragile output pattern matching.
---
## API Surface
The repo exposes CLI-level APIs rather than import-oriented libraries.
### Shell API
`fleet-launch.sh`
- `./fleet-launch.sh mac`
- `./fleet-launch.sh allegro`
- `./fleet-launch.sh both`
### Python CLIs
`fleet-christen.py`
- `python3 fleet-christen.py mac`
- `python3 fleet-christen.py allegro`
- `python3 fleet-christen.py both`
`fleet-dispatch.py`
- `python3 fleet-dispatch.py`
- `python3 fleet-dispatch.py --cycles 10 --interval 60`
- `python3 fleet-dispatch.py --machine mac`
`fleet-status.py`
- `python3 fleet-status.py`
- `python3 fleet-status.py --machine allegro`
### Internal function surface worth naming explicitly
`fleet-launch.sh`
- `parse_spec()`
- `launch_local()`
- `launch_remote()`
`fleet-christen.py`
- `send_keys()`
- `christen_window()`
- `christen_machine()`
- `christen_remote()`
`fleet-dispatch.py`
- `load_token()`
- `load_spec()`
- `load_state()`
- `save_state()`
- `get_issues()`
- `send_to_pane()`
- `comment_on_issue()`
- `build_prompt()`
- `dispatch_cycle()`
- `dispatch_council()`
`fleet-status.py`
- `get_pane_status()`
- `check_machine()`
These are the true API surface for future hardening and testing.
---
## Test Coverage Gaps
### Current state
Grounded from the pipeline dry run on `/tmp/burn-fleet-genome`:
- 0% estimated coverage
- untested modules called out by pipeline: `fleet-christen`, `fleet-dispatch`, `fleet-status`
- no checked-in automated test suite
### Critical paths with no tests
1. `send_to_pane()`
- local tmux command construction
- remote SSH command construction
- escaping of issue titles and prompts
- failure handling when tmux or SSH fails
2. `comment_on_issue()`
- verifies Gitea comment formatting
- verifies non-200 responses do not silently disappear
3. `get_issues()`
- PR filtering
- `SKIP_LABELS` filtering
- title-based meta filtering
- robustness when Gitea returns malformed or partial issue objects
4. `dispatch_cycle()`
- correct pooling by window
- deduplication via `dispatch-state.json`
- pane recycling behavior
- correctness when one repo has zero issues and another has many
5. `get_pane_status()`
- classification heuristics for working/idle/error/dead
- false positives from incidental strings like `error` in normal output
6. `fleet-launch.sh`
- parse correctness for pane counts
- layout creation behavior across first vs later windows
- remote script generation for Allegro
### Missing tests to generate next in the real target repo
If the goal is to harden `burn-fleet` itself, the first tests to add should be:
- `test_route_tables_cover_spec_windows`
- `test_send_to_pane_escapes_single_quotes_and_special_chars`
- `test_comment_on_issue_formats_machine_window_pane_body`
- `test_get_issues_skips_prs_and_meta_labels`
- `test_dispatch_cycle_persists_dispatch_state_once`
- `test_get_pane_status_classifies_spinner_vs_traceback_vs_empty`
These are the minimum critical-path tests.
---
## Security Considerations
### 1. Command injection surface
`send_to_pane()` and the remote tmux/SSH command assembly are the biggest security surface.
Even though single quotes are escaped in prompts, this remains a command injection boundary because untrusted issue titles and repo metadata cross into shell commands.
This is why `command injection` is the right risk label for the repo.
The risk is not hypothetical; the repo is literally translating issue text into shell transport.
### 2. Credential handling
The dispatcher uses a local token file for Gitea authentication.
That is a credential handling concern because:
- token locality is assumed
- file path and host assumptions are embedded into runtime code
- there is no retry / fallback / explicit missing-token UX beyond failure
### 3. SSH trust boundary
Remote pane control over `root@167.99.126.228` means the repo assumes a trusted SSH path to a root shell.
That is operationally powerful and dangerous.
A malformed remote command, stale known_hosts state, or wrong host mapping has fleet-wide consequences.
### 4. Runtime state tampering
`dispatch-state.json` is a local mutable state file with no locking, signing, or cross-machine reconciliation.
If it is corrupted or lost, deduplication semantics fail.
That can cause repeated dispatches or misleading status.
### 5. Live-forge mutation
`comment_on_issue()` mutates live issue threads on every dispatch cycle.
That means any bug in deduplication or routing will create visible comment spam on the forge.
### 6. Dependency risk
The repo depends on `requests` for Gitea API access but has no pinned dependency metadata or environment contract in-repo.
This is a small operational repo, but reproducibility is weak.
---
## Dependency Picture
### Runtime dependencies
- Python 3
- `requests`
- tmux
- SSH client
- ssh trust boundary to `root@167.99.126.228`
- access to a Gitea token file
### Implied environment dependencies
- active tmux sessions on Mac and Allegro
- SSH trust / connectivity to the VPS
- hermes available in pane environments
- Gitea reachable at `https://forge.alexanderwhitestone.com`
### Notably missing
- no `requirements.txt`
- no `pyproject.toml`
- no explicit test harness
- no schema validation for `fleet-spec.json`
---
## Performance Characteristics
For such a small repo, the performance question is not CPU time inside Python.
It is orchestration fan-out latency.
The main scaling costs are:
- repeated Gitea issue fetches across repos
- SSH round-trips to Allegro
- tmux pane fan-out across 100+ panes
- serialized `time.sleep(0.2)` dispatch staggering
This means the bottleneck is control-plane coordination, not computation.
The repo will scale until SSH / tmux / Gitea latency become dominant.
---
## Dead Code / Drift Risks
### 1. Spec vs route duplication
`fleet-spec.json` defines windows and lanes, while `fleet-dispatch.py` separately defines `MAC_ROUTE` and `ALLEGRO_ROUTE`.
That is the biggest drift risk.
A window can exist in the spec and be missing from a route table, or vice versa.
### 2. Runtime-generated files absent from repo contracts
`dispatch-state.json` is operationally critical but not described as a first-class contract in code.
The repo assumes it exists or can be created, but does not validate structure.
### 3. README drift risk
The README says "use fleet-christen.sh" in one place while the actual file is `fleet-christen.py`.
That is a small but real operator-footgun and a sign the human runbook can drift from the executable surface.
---
## Suggested Follow-up Work
1. Move repo-to-window routing into `fleet-spec.json` and derive `MAC_ROUTE` / `ALLEGRO_ROUTE` programmatically.
2. Add automated tests for `send_to_pane`, `get_issues`, `dispatch_cycle`, and `get_pane_status`.
3. Add a schema validator for `fleet-spec.json`.
4. Add explicit dependency metadata (`requirements.txt` or `pyproject.toml`).
5. Add dry-run / no-side-effect mode for dispatch and christening.
6. Add retry/backoff and error reporting around Gitea comments and SSH execution.
---
## Bottom Line
`burn-fleet` is a small repo with outsized operational leverage.
Its genome is simple:
- one declarative topology file
- four operational adapters
- one local runtime ledger
- many side effects across tmux, SSH, and Gitea
It already expresses the philosophy of narrative-driven infrastructure well.
What it lacks is not architecture.
What it lacks is hardening:
- tests around the dangerous paths
- centralization of duplicated routing truth
- stronger command / credential / runtime-state safeguards
That makes it a strong control-plane prototype and a weakly tested production surface.

View File

@@ -0,0 +1,101 @@
# GENOME.md — Burn Fleet (Timmy_Foundation/burn-fleet)
> Codebase Genome v1.0 | Generated 2026-04-16 | Repo 14/16
## Project Overview
**Burn Fleet** is the autonomous dispatch infrastructure for the Timmy Foundation. It manages 112 tmux panes across Mac and VPS, routing Gitea issues to lane-specialized workers by repo. Each agent has a mythological name — they are all Timmy with different hats.
**Core principle:** Dispatch ALL panes. Never scan for idle. Stale work beats idle workers.
## Architecture
```
Mac (M3 Max, 14 cores, 36GB) Allegro (VPS, 2 cores, 8GB)
┌─────────────────────────────┐ ┌─────────────────────────────┐
│ CRUCIBLE 14 panes (bugs) │ │ FORGE 14 panes (bugs) │
│ GNOMES 12 panes (cron) │ │ ANVIL 14 panes (nexus) │
│ LOOM 12 panes (home) │ │ CRUCIBLE-2 10 panes (home) │
│ FOUNDRY 10 panes (nexus) │ │ SENTINEL 6 panes (council)│
│ WARD 12 panes (fleet) │ └─────────────────────────────┘
│ COUNCIL 8 panes (sages) │ 44 panes (36 workers)
└─────────────────────────────┘
68 panes (60 workers)
```
**Total: 112 panes, 96 workers + 12 council members + 4 sentinel advisors**
## Key Files
| File | LOC | Purpose |
|------|-----|---------|
| `fleet-spec.json` | ~200 | Machine definitions, window layouts, lane assignments, agent names |
| `fleet-launch.sh` | ~100 | Create tmux sessions with correct pane counts on Mac + Allegro |
| `fleet-christen.py` | ~80 | Launch hermes in all panes and send identity messages |
| `fleet-dispatch.py` | ~250 | Pull Gitea issues and route to correct panes by lane |
| `fleet-status.py` | ~100 | Health check across all machines |
| `allegro/docker-compose.yml` | ~30 | Allegro VPS container definition |
| `allegro/Dockerfile` | ~20 | Allegro build definition |
| `allegro/healthcheck.py` | ~15 | Allegro container health check |
**Total: ~800 LOC**
## Lane Routing
Issues are routed by repo to the correct window:
| Repo | Mac Window | Allegro Window |
|------|-----------|----------------|
| hermes-agent | CRUCIBLE, GNOMES | FORGE |
| timmy-home | LOOM | CRUCIBLE-2 |
| timmy-config | LOOM | CRUCIBLE-2 |
| the-nexus | FOUNDRY | ANVIL |
| the-playground | — | ANVIL |
| the-door | WARD | CRUCIBLE-2 |
| fleet-ops | WARD | CRUCIBLE-2 |
| turboquant | WARD | — |
## Entry Points
| Command | Purpose |
|---------|---------|
| `./fleet-launch.sh both` | Create tmux layout on Mac + Allegro |
| `python3 fleet-christen.py both` | Wake all agents with identity messages |
| `python3 fleet-dispatch.py --cycles 1` | Single dispatch cycle |
| `python3 fleet-dispatch.py --cycles 10 --interval 60` | Continuous burn (10 cycles, 60s apart) |
| `python3 fleet-status.py` | Health check all machines |
## Agent Names
| Window | Names | Count |
|--------|-------|-------|
| CRUCIBLE | AZOTH, ALBEDO, CITRINITAS, RUBEDO, SULPHUR, MERCURIUS, SAL, ATHANOR, VITRIOL, SATURN, JUPITER, MARS, EARTH, SOL | 14 |
| GNOMES | RAZIEL, AZRAEL, CASSIEL, METATRON, SANDALPHON, BINAH, CHOKMAH, KETER, ALDEBARAN, RIGEL, SIRIUS, POLARIS | 12 |
| FORGE | HAMMER, ANVIL, ADZE, PICK, TONGS, WRENCH, SCREWDRIVER, BOLT, SAW, TRAP, HOOK, MAGNET, SPARK, FLAME | 14 |
| COUNCIL | TESLA, HERMES, GANDALF, DAVINCI, ARCHIMEDES, TURING, AURELIUS, SOLOMON | 8 |
## Design Decisions
1. **Separate GILs** — Allegro runs Python independently on VPS for true parallelism
2. **Queue, not send-keys** — Workers process at their own pace, no interruption
3. **Lane enforcement** — Panes stay in one repo to build deep context
4. **Dispatch ALL panes** — Never scan for idle; stale work beats idle workers
5. **Council is advisory** — Named archetypes provide perspective, not task execution
## Scaling
- Add panes: Edit `fleet-spec.json``fleet-launch.sh``fleet-christen.py`
- Add machines: Edit `fleet-spec.json` → Add routing in `fleet-dispatch.py` → Ensure SSH access
## Sovereignty Assessment
- **Fully local** — Mac + user-controlled VPS, no cloud dependencies
- **No phone-home** — Gitea API is self-hosted
- **Open source** — All code on Gitea
- **SSH-based** — Mac → Allegro communication via SSH only
**Verdict: Fully sovereign. Autonomous fleet dispatch with no external dependencies.**
---
*"Dispatch ALL panes. Never scan for idle — stale work beats idle workers."*

View File

@@ -0,0 +1,137 @@
# GENOME.md — Evennia Local World (Timmy_Foundation/the-nexus → evennia_mempalace)
> Codebase Genome v1.0 | Generated 2026-04-15 | Repo 10/16
## Project Overview
**Evennia Local World** is the MUD (Multi-User Dungeon) layer of the sovereign fleet. Implemented as a subsystem within `the-nexus`, it provides a persistent text-based world where Timmy's agents can navigate rooms, interact with NPCs, and access the MemPalace memory system through traditional MUD commands.
**Core principle:** Evennia owns persistent world truth. Nexus owns visualization. The adapter owns only translation.
## Architecture
```mermaid
graph TD
subgraph "Evennia MUD World"
ROOMS[MemPalaceRoom Typeclasses]
NPCS[AI NPCs]
COMMANDS[recall / write commands]
end
subgraph "Event Adapter"
EA[nexus/evennia_event_adapter.py]
WS[nexus/evennia_ws_bridge.py]
end
subgraph "Nexus Visualization"
THREE[Three.js 3D World]
SESSIONS[session-rooms.js]
end
subgraph "MemPalace Memory"
SEARCH[nexus/mempalace/searcher.py]
CONFIG[nexus/mempalace/config.py]
end
ROOMS --> SEARCH
COMMANDS --> SEARCH
ROOMS --> EA
NPCS --> EA
EA --> WS
WS --> THREE
WS --> SESSIONS
```
## Key Components
| Component | Path | Purpose |
|-----------|------|---------|
| MemPalaceRoom | `nexus/evennia_mempalace/typeclasses/rooms.py` | Room typeclass backed by live MemPalace search |
| AI NPCs | `nexus/evennia_mempalace/typeclasses/npcs.py` | NPCs with AI personality and memory |
| recall command | `nexus/evennia_mempalace/commands/recall.py` | Search MemPalace from within MUD |
| write command | `nexus/evennia_mempalace/commands/write.py` | Record artifacts to MemPalace |
| Event Adapter | `nexus/evennia_event_adapter.py` | Evennia → Nexus event translation |
| WS Bridge | `nexus/evennia_ws_bridge.py` | WebSocket bridge for real-time sync |
| Multi-User Bridge | `world/multi_user_bridge.py` | Multi-user session management |
## Event Protocol
The Evennia → Nexus event protocol defines canonical event families:
| Event Type | Purpose |
|------------|---------|
| `evennia.session_bound` | Binds Hermes session to world interaction |
| `evennia.actor_located` | Declares current location |
| `evennia.room_described` | Room description rendered |
| `evennia.command_executed` | MUD command processed |
| `evennia.memory_recalled` | MemPalace search result |
## Room Types
| Type | Description |
|------|-------------|
| MemPalaceRoom | Description auto-refreshes from live palace search |
| Standard rooms | Static descriptions from world config |
Room descriptions update on entry via `search_memories(room_topic)` from `nexus.mempalace.searcher`.
## MemPalace Room Taxonomy
The MUD world maps to the fleet's MemPalace taxonomy:
```
WING: [wizard_name]
ROOM: forge — CI, builds, infrastructure
ROOM: hermes — Agent platform, harness
ROOM: nexus — Reports, documentation
ROOM: issues — Tickets, PR summaries
ROOM: experiments — Spikes, prototypes
ROOM: sovereign — Alexander's requests & responses
```
Each room is a `MemPalaceRoom` typeclass that pulls live content from the palace.
## Commands
| Command | File | Purpose |
|---------|------|---------|
| `recall <query>` | commands/recall.py | Search MemPalace from MUD |
| `write <room> <text>` | commands/write.py | Record artifact to MemPalace |
## Test Coverage
| Test | File | Validates |
|------|------|-----------|
| Event adapter | tests/test_evennia_event_adapter.py | Event translation |
| Mempalace commands | tests/test_evennia_mempalace_commands.py | recall/write commands |
| WS bridge | tests/test_evennia_ws_bridge.py | WebSocket communication |
| Room validation | tests/test_mempalace_validate_rooms.py | Room taxonomy compliance |
## File Index
| File | Purpose |
|------|---------|
| `nexus/evennia_mempalace/__init__.py` | Package init |
| `nexus/evennia_mempalace/typeclasses/rooms.py` | MemPalaceRoom typeclass |
| `nexus/evennia_mempalace/typeclasses/npcs.py` | AI NPC typeclasses |
| `nexus/evennia_mempalace/commands/recall.py` | recall command |
| `nexus/evennia_mempalace/commands/write.py` | write command |
| `nexus/evennia_event_adapter.py` | Event protocol adapter |
| `nexus/evennia_ws_bridge.py` | WebSocket bridge |
| `world/multi_user_bridge.py` | Multi-user session bridge |
| `EVENNIA_NEXUS_EVENT_PROTOCOL.md` | Protocol specification |
| `FIRST_LIGHT_REPORT_EVENNIA_BRIDGE.md` | Initial deployment report |
## Sovereignty Assessment
- **Fully local** — Evennia runs on the user's machine or sovereign VPS
- **No phone-home** — All communication is user-controlled WebSocket
- **Open source** — Evennia 6.0 is MIT licensed
- **Fleet-integrated** — Direct MemPalace access via recall/write commands
- **Multi-user** — Supports multiple simultaneous players
**Verdict: Fully sovereign. Persistent text-based world with AI memory integration.**
---
*"Evennia owns persistent world truth. Nexus owns visualization. The adapter owns only translation, not storage or game logic."*

397
genomes/fleet-ops-GENOME.md Normal file
View File

@@ -0,0 +1,397 @@
# GENOME.md — fleet-ops
Host artifact for timmy-home issue #680. The analyzed code lives in the separate `fleet-ops` repository; this document is the curated genome written from a fresh clone of that repo at commit `38c4eab`.
## Project Overview
`fleet-ops` is the infrastructure and operations control plane for the Timmy Foundation fleet. It is not a single deployable application. It is a mixed ops repository with four overlapping layers:
1. Ansible orchestration for VPS provisioning and service rollout.
2. Small Python microservices for shared fleet state.
3. Cron- and CLI-driven operator scripts.
4. A separate local `docker-compose.yml` sandbox for a simplified all-in-one stack.
Two facts shape the repo more than anything else:
- The real fleet deployment path starts at `site.yml``playbooks/site.yml` and lands services through Ansible roles.
- The repo also contains several aspirational or partially wired Python modules whose names imply runtime importance but whose deployment path is weak, indirect, or missing.
Grounded metrics from the fresh analysis run:
- `python3 ~/.hermes/pipelines/codebase-genome.py --path /tmp/fleet-ops-genome --dry-run` reported `97` source files, `12` test files, `29` config files, and `16,658` total lines.
- A local filesystem count found `39` Python source files, `12` Python test files, and `74` YAML files.
- `python3 -m pytest -q --continue-on-collection-errors` produced `158 passed, 1 failed, 2 errors`.
The repo is therefore operationally substantial, but only part of that surface is coherently tested and wired.
## Architecture
```mermaid
graph TD
A[site.yml] --> B[playbooks/site.yml]
B --> C[preflight.yml]
B --> D[baseline.yml]
B --> E[deploy_ollama.yml]
B --> F[deploy_gitea.yml]
B --> G[deploy_hermes.yml]
B --> H[deploy_conduit.yml]
B --> I[harmony_audit role]
G --> J[playbooks/host_vars/* wizard_instances]
G --> K[hermes-agent role]
K --> L[systemd wizard services]
M[templates/fleet-deploy-hook.service] --> N[scripts/deploy-hook.py]
N --> B
O[playbooks/roles/message-bus/templates/busd.service.j2] --> P[message_bus.py]
Q[playbooks/roles/knowledge-store/templates/knowledged.service.j2] --> R[knowledge_store.py]
S[registry.yaml] --> T[health_dashboard.py]
S --> U[scripts/registry_health_updater.py]
S --> V[federation_sync.py]
W[cron/dispatch-consumer.yml] --> X[scripts/dispatch_consumer.py]
Y[morning_report_cron.yml] --> Z[scripts/morning_report_compile.py]
AA[nightly_efficiency_cron.yml] --> AB[scripts/nightly_efficiency_report.py]
AC[burndown_watcher_cron.yml] --> AD[scripts/burndown_cron.py]
AE[docker-compose.yml] --> AF[local ollama]
AE --> AG[local gitea]
AE --> AH[agent container]
AE --> AI[monitor loop]
```
### Structural read
The cleanest mental model is not “one app,” but “one repo that tries to be the fleets operator handbook, deployment engine, shared service shelf, and scratchpad.”
That produces three distinct control planes:
1. `playbooks/` is the strongest source of truth for VPS deployment.
2. `registry.yaml` and `manifest.yaml` act as runtime or operator registries for scripts.
3. `docker-compose.yml` models a separate local sandbox whose assumptions do not fully match the Ansible path.
## Entry Points
### Primary fleet deploy entry points
- `site.yml` — thin repo-root wrapper that imports `playbooks/site.yml`.
- `playbooks/site.yml` — multi-phase orchestrator for preflight, baseline, Ollama, Gitea, Hermes, Conduit, and local harmony audit.
- `playbooks/deploy_hermes.yml` — the most important service rollout for wizard instances; requires `wizard_instances` and pulls `vault_openrouter_api_key` / `vault_openai_api_key`.
- `playbooks/provision_and_deploy.yml` — DigitalOcean create-and-bootstrap path using `community.digital.digital_ocean_droplet` and a dynamic `new_droplets` group.
### Deployed service entry points
- `message_bus.py` — HTTP message queue service deployed by `playbooks/roles/message-bus/templates/busd.service.j2`.
- `knowledge_store.py` — SQLite-backed shared fact service deployed by `playbooks/roles/knowledge-store/templates/knowledged.service.j2`.
- `scripts/deploy-hook.py` — webhook listener launched by `templates/fleet-deploy-hook.service` with `ExecStart=/usr/bin/python3 /opt/fleet-ops/scripts/deploy-hook.py`.
### Cron and operator entry points
- `scripts/dispatch_consumer.py` — wired by `cron/dispatch-consumer.yml`.
- `scripts/morning_report_compile.py` — wired by `morning_report_cron.yml`.
- `scripts/nightly_efficiency_report.py` — wired by `nightly_efficiency_cron.yml`.
- `scripts/burndown_cron.py` — wired by `burndown_watcher_cron.yml`.
- `scripts/fleet_readiness.py` — operator validation script for `manifest.yaml`.
- `scripts/fleet-status.py` — prints a fleet status snapshot directly from top-level code.
### CI / verification entry points
- `.gitea/workflows/ansible-lint.yml` — YAML lint, `ansible-lint`, syntax checks, inventory validation.
- `.gitea/workflows/auto-review.yml` — lightweight review workflow with YAML lint, syntax checks, secret scan, and merge-conflict probe.
### Local development stack entry point
- `docker-compose.yml` — brings up `ollama`, `gitea`, `agent`, and `monitor` for a local stack.
## Data Flow
### 1) Deploy path
1. A repo operator pushes or references deployable state.
2. `scripts/deploy-hook.py` receives the webhook.
3. The hook updates `/opt/fleet-ops`, then invokes Ansible.
4. `playbooks/site.yml` fans into phase playbooks.
5. `playbooks/deploy_hermes.yml` renders per-instance config and systemd services from `wizard_instances` in `playbooks/host_vars/*`.
6. Services expose local `/health` endpoints on assigned ports.
### 2) Shared service path
1. Agents or tools post work to `message_bus.py`.
2. Consumers poll `/messages` and inspect `/queue`, `/deadletter`, and `/audit`.
3. Facts are written into `knowledge_store.py` and federated through peer sync endpoints.
4. `health_dashboard.py` and `scripts/registry_health_updater.py` read `registry.yaml` and probe service URLs.
### 3) Reporting path
1. Cron YAML launches queue/report scripts.
2. Scripts read `~/.hermes/`, Gitea APIs, local logs, or registry files.
3. Output is emitted as JSON, markdown, or console summaries.
### Important integration fracture
`federation_sync.py` does not currently match the services it tries to coordinate.
- `message_bus.py` returns `/messages` as `{"messages": [...], "count": N}` at line 234.
- `federation_sync.py` polls `.../messages?limit=50` and then only iterates if `isinstance(data, list)` at lines 136-140.
- `federation_sync.py` also requests `.../knowledge/stats` at line 230, but `knowledge_store.py` documents `/sync/status`, `/facts`, and `/peers`, not `/knowledge/stats`.
This means the repo contains a federation layer whose assumed contracts drift from the concrete microservices beside it.
## Key Abstractions
### `MessageStore` in `message_bus.py`
Core in-memory queue abstraction. It underlies:
- enqueue / poll behavior
- TTL expiry and dead-letter handling
- queue stats and audit trail endpoints
The tests in `tests/test_message_bus.py` make this one of the best-specified components in the repo.
### `KnowledgeDB` in `knowledge_store.py`
SQLite-backed fact registry with HTTP exposure for:
- storing facts
- querying and deleting facts
- peer registration
- push/pull federation
- sync status reporting
This is the nearest thing the repo has to a durable shared memory service.
### `FleetMonitor` in `health_dashboard.py`
Loads `registry.yaml`, polls wizard endpoints, caches results, and exposes both HTML and JSON views. It is the operator-facing read model of the fleet.
### `SyncEngine` in `federation_sync.py`
Intended as the bridge across message bus, audit trail, and knowledge store. The design intent is strong, but the live endpoint contracts appear out of sync.
### `ProfilePolicy` in `scripts/profile_isolation.py`
Encodes tmux/agent lifecycle policy by profile. This is one of the more disciplined “ops logic” modules: focused, testable, and bounded.
### `GenerationResult` / `VideoEngineClient` in `scripts/video_engine_client.py`
Represents the repos media-generation sidecar boundary. The code is small and clear, but its tests are partially stale relative to implementation behavior.
## API Surface
### `message_bus.py`
Observed HTTP surface includes:
- `POST /message`
- `GET /messages?to=<agent>&limit=<n>`
- `GET /queue`
- `GET /deadletter`
- `GET /audit`
- `GET /health`
### `knowledge_store.py`
Documented surface includes:
- `POST /fact`
- `GET /facts`
- `DELETE /facts/<key>`
- `POST /sync/pull`
- `POST /sync/push`
- `GET /sync/status`
- `GET /peers`
- `POST /peers`
- `GET /health`
### `health_dashboard.py`
- `/`
- `/api/status`
- `/api/wizard/<id>`
### `scripts/deploy-hook.py`
- `/health`
- `/webhook`
### Ansible operator surface
Primary commands implied by the repo:
- `ansible-playbook -i playbooks/inventory site.yml`
- `ansible-playbook -i playbooks/inventory playbooks/provision_and_deploy.yml`
- `ansible-playbook -i playbooks/inventory playbooks/deploy_hermes.yml`
## Dependencies
### Python and shell posture
The repo is mostly Python stdlib plus Ansible/shell orchestration. It is not packaged as a single installable Python project.
### Explicit Ansible collections
`requirements.yml` declares:
- `community.docker`
- `community.general`
- `ansible.posix`
The provisioning docs and playbooks also rely on `community.digital.digital_ocean_droplet` in `playbooks/provision_and_deploy.yml`.
### External service dependencies
- Gitea
- Ollama
- DigitalOcean
- systemd
- Docker / Docker Compose
- local `~/.hermes/` session and burn-log state
### Hidden runtime dependency
Several conceptual modules import `hermes_tools` directly:
- `compassion_layer.py`
- `sovereign_librarian.py`
- `sovereign_muse.py`
- `sovereign_pulse.py`
- `sovereign_sentinel.py`
- `synthesis_engine.py`
That dependency is not self-contained inside the repo and directly causes the local collection errors.
## Test Coverage Gaps
### Current tested strengths
The strongest, most trustworthy tests are around:
- `tests/test_message_bus.py`
- `tests/test_knowledge_store.py`
- `tests/test_health_dashboard.py`
- `tests/test_registry_health_updater.py`
- `tests/test_profile_isolation.py`
- `tests/test_skill_scorer.py`
- `tests/test_nightly_efficiency_report.py`
Those files make the shared-service core much more legible than the deployment layer.
### Current local status
Fresh run result:
- `158 passed, 1 failed, 2 errors`
Collection errors:
- `tests/test_heart.py` fails because `compassion_layer.py` imports `hermes_tools`.
- `tests/test_synthesis.py` fails because `sovereign_librarian.py` imports `hermes_tools`.
Runnable failure:
- `tests/test_video_engine_client.py` expects `generate_draft()` to raise on HTTP 503.
- `scripts/video_engine_client.py` currently catches exceptions and returns `GenerationResult(success=False, error=...)` instead.
### High-value untested paths
The most important missing or weakly validated surfaces are:
- `scripts/deploy-hook.py` — high-blast-radius deploy trigger.
- `playbooks/deploy_gitea.yml` / `playbooks/deploy_hermes.yml` / `playbooks/provision_and_deploy.yml` — critical control plane, almost entirely untested in-repo.
- `scripts/morning_report_compile.py` — cron-facing reporting logic.
- `scripts/burndown_cron.py` and related watcher scripts.
- `scripts/generate_video.py`, `scripts/tiered_render.py`, and broader video-engine operator paths.
- `scripts/fleet-status.py` — prints directly from module scope and has no `__main__` guard.
### Coverage quality note
The repos best tests cluster around internal Python helpers. The repos biggest operational risk lives in deployment, cron wiring, and shell/Ansible behaviors that are not equivalently exercised.
## Security Considerations
### Strong points
- Vault use exists in `playbooks/group_vars/vault.yml` and inline vaulted material in `manifest.yaml`.
- `playbooks/deploy_gitea.yml` sets `gitea_disable_registration: true`, `gitea_require_signin: true`, and `gitea_register_act_runner: false`.
- The Hermes role renders per-instance env/config and uses systemd hardening patterns.
- Gitea, Nostr relay, and other web surfaces are designed around nginx/TLS roles.
### Concrete risks
1. `scripts/deploy-hook.py` explicitly disables signature enforcement when `DEPLOY_HOOK_SECRET` is unset.
2. `playbooks/roles/gitea/defaults/main.yml` sets `gitea_webhook_allowed_host_list: "*"`.
3. Both `ansible.cfg` files disable host key checking.
4. The repo has multiple sources of truth for ports and service topology:
- `playbooks/host_vars/ezra-primary.yml` uses `8643`
- `manifest.yaml` uses `8643`
- `registry.yaml` points Ezra health to `8646`
5. `registry.yaml` advertises services like `busd`, `auditd`, and `knowledged`, but the main `playbooks/site.yml` phases do not include message-bus or knowledge-store roles.
### Drift / correctness risks that become security risks
- `playbooks/deploy_auto_merge.yml` targets `hosts: gitea_servers`, but the inventory groups visible in `playbooks/inventory` are `forge`, `vps`, `agents`, and `wizards`.
- `playbooks/roles/gitea/defaults/main.yml` includes runner labels with a probable typo: `ubuntu-22.04:docker://catthehocker/ubuntu:act-22.04`.
- The local compose quick start is not turnkey: `Dockerfile.agent` copies `requirements-agent.txt*` and `agent/`, but the runtime falls back to a tiny health/tick loop if the real agent source is absent.
## Deployment
### VPS / real fleet path
Repo-root wrapper:
```bash
ansible-playbook -i playbooks/inventory site.yml
```
Direct orchestrator:
```bash
ansible-playbook -i playbooks/inventory playbooks/site.yml
```
Provision and bootstrap a new node:
```bash
ansible-playbook -i playbooks/inventory playbooks/provision_and_deploy.yml
```
### Local sandbox path
```bash
cp .env.example .env
docker compose up -d
```
But this path must be read skeptically. `docker-compose.yml` is a local convenience stack, while the real fleet path uses Ansible + systemd + host vars + vault-backed secrets.
## Dead Code Candidates and Operator Footguns
- `scripts/fleet-status.py` behaves like a one-shot report script with top-level execution, not a reusable CLI module.
- `README.md` ends with a visibly corrupted Nexus Watchdog section containing broken formatting.
- `Sovereign_Health_Check.md` still recommends running the broken `tests/test_heart.py` and `tests/test_synthesis.py` health suite.
- `federation_sync.py` currently looks architecturally important but contractually out of sync with `message_bus.py` and `knowledge_store.py`.
## Bottom Line
`fleet-ops` contains the real bones of a sovereign fleet control plane, but those bones are unevenly ossified.
The strong parts are:
- the phase-based Ansible deployment structure in `playbooks/site.yml`
- the microservice-style core in `message_bus.py`, `knowledge_store.py`, and `health_dashboard.py`
- several focused Python test suites that genuinely specify behavior
The weak parts are:
- duplicated sources of truth (`playbooks/host_vars/*`, `manifest.yaml`, `registry.yaml`, local compose)
- deployment and cron surfaces that matter more operationally than they are tested
- conceptual “sovereign_*” modules that pull in `hermes_tools` and currently break local collection
If this repo were being hardened next, the highest-leverage moves would be:
1. Make the registries consistent (`8643` vs `8646`, service inventory vs deployed phases).
2. Add focused tests around `scripts/deploy-hook.py` and the deploy/report cron scripts.
3. Decide which Python modules are truly production runtime and which are prototypes, then wire or prune accordingly.
4. Collapse the number of “truth” files an operator has to trust during a deploy.

160
genomes/the-nexus/GENOME.md Normal file
View File

@@ -0,0 +1,160 @@
# GENOME.md — The Nexus (Timmy_Foundation/the-nexus)
> Codebase Genome v1.0 | Generated 2026-04-15 | Repo 5/16
## Project Overview
**The Nexus** is a dual-purpose project: a local-first training ground for Timmy AI agents and a wizardly visualization surface for the sovereign fleet. It combines a Three.js 3D world, Evennia MUD integration, MemPalace memory system, and fleet intelligence infrastructure.
**Core principle:** agents work, the world visualizes, memory persists.
## Architecture
```mermaid
graph TD
subgraph "3D World (Three.js)"
APP[app.js] --> SCENE[Scene Manager]
SCENE --> PORTALS[Portal System]
SCENE --> PARTICLES[Particle Engine]
SCENE --> MEMPALACE_3D[MemPalace 3D]
end
subgraph "Backend (Python)"
SERVER[server.py] --> NEXUS[nexus/]
NEXUS --> MEMPALACE[mempalace/]
NEXUS --> FLEET[fleet/]
NEXUS --> AGENT[agent/]
NEXUS --> INTEL[intelligence/]
end
subgraph "Evennia MUD Bridge"
NEXUS --> EVENNIA[nexus/evennia_mempalace/]
EVENNIA --> ROOMS[Room Typeclasses]
EVENNIA --> COMMANDS[Recall/Write Commands]
end
subgraph "Build & Deploy"
DOCKER[docker-compose.yml] --> SERVER
DEPLOY[deploy.sh] --> VPS[VPS Deployment]
end
```
## Key Subsystems
| Subsystem | Path | Purpose |
|-----------|------|---------|
| Three.js 3D World | `app.js`, `index.html` | Browser-based 3D visualization surface |
| Portal System | `portals.json`, commands/ | Teleportation between world zones |
| MemPalace | `mempalace/`, `nexus/mempalace/` | Fleet memory: rooms, search, retention |
| Evennia Bridge | `nexus/evennia_mempalace/` | MUD world ↔ MemPalace integration |
| Fleet Intelligence | `fleet/`, `intelligence/` | Cross-wizard analytics and coordination |
| Agent Tools | `agent/` | Agent capabilities and tool definitions |
| Boot System | `boot.js`, `bootstrap.mjs` | World initialization and startup |
| Evolution | `evolution/` | System evolution tracking and proposals |
| GOFAI Worker | `gofai_worker.js` | Classical AI logic engine |
| Concept Packs | `concept-packs/` | World content and knowledge packs |
| Gitea Integration | `gitea_api/` | Forge API helpers and automation |
## Entry Points
| Entry Point | File | Purpose |
|-------------|------|---------|
| Browser | `index.html` | Three.js 3D world entry |
| Node Server | `server.py` | Backend API and WebSocket server |
| Electron | `electron-main.js` | Desktop app shell |
| Deploy | `deploy.sh` | VPS deployment script |
| Docker | `docker-compose.yml` | Containerized deployment |
## MemPalace System
The MemPalace is the fleet's persistent memory:
- **Rooms:** forge, hermes, nexus, issues, experiments (core) + optional domain rooms
- **Taxonomy:** Defined in `mempalace/rooms.yaml` (fleet standard)
- **Search:** `nexus/mempalace/searcher.py` — semantic search across rooms
- **Fleet API:** `mempalace/fleet_api.py` — HTTP API for cross-wizard memory access
- **Retention:** `mempalace/retain_closets.py` — 90-day auto-pruning
- **Tunnel Sync:** `mempalace/tunnel_sync.py` — Cross-wing room synchronization
- **Privacy Audit:** `mempalace/audit_privacy.py` — Data privacy compliance
## Evennia Integration
The Evennia bridge connects the 3D world to a traditional MUD:
- **Room Typeclasses:** `nexus/evennia_mempalace/typeclasses/rooms.py` — MemPalace-aware rooms
- **NPCs:** `nexus/evennia_mempalace/typeclasses/npcs.py` — AI-powered NPCs
- **Commands:** `nexus/evennia_mempalace/commands/` — recall, write, and exploration commands
- **Protocol:** `EVENNIA_NEXUS_EVENT_PROTOCOL.md` — Event bridge specification
## Configuration
| File | Purpose |
|------|---------|
| `config/` | World configuration |
| `portals.json` | Portal definitions and teleportation |
| `vision.json` | Visual rendering configuration |
| `docker-compose.yml` | Container orchestration |
| `Dockerfile` | Build definition |
## Test Coverage
| Area | Tests | Notes |
|------|-------|-------|
| CI Workflows | `.gitea/workflows/`, `.github/` | Smoke tests, linting |
| Python | Limited | Core nexus modules lack unit tests |
| JavaScript | Limited | No dedicated test suite for 3D world |
| Integration | Manual | Evennia bridge tested via telnet |
## Documentation
| File | Purpose |
|------|---------|
| `README.md` | Branch protection policy + project overview |
| `DEVELOPMENT.md` | Dev setup guide |
| `CONTRIBUTING.md` | Contribution guidelines |
| `SOUL.md` | Project values and philosophy |
| `POLICY.md` | Operational policies |
| `EVENNIA_NEXUS_EVENT_PROTOCOL.md` | Evennia bridge spec |
| `GAMEPORTAL_PROTOCOL.md` | Game portal specification |
| `FIRST_LIGHT_REPORT.md` | Initial deployment report |
| `docs/` | Extended documentation |
## File Structure (Top Level)
```
the-nexus/
├── app.js # Three.js application
├── index.html # Browser entry point
├── server.py # Backend server
├── boot.js # Boot sequence
├── bootstrap.mjs # ES module bootstrap
├── electron-main.js # Desktop app
├── deploy.sh # VPS deployment
├── docker-compose.yml # Container config
├── nexus/ # Python core modules
│ ├── evennia_mempalace/ # Evennia MUD bridge
│ └── mempalace/ # Memory system
├── mempalace/ # Fleet memory tools
├── fleet/ # Fleet coordination
├── agent/ # Agent tools
├── intelligence/ # Cross-wizard analytics
├── commands/ # World commands
├── concept-packs/ # Content packs
├── evolution/ # System evolution
├── assets/ # Static assets
└── docs/ # Documentation
```
## Sovereignty Assessment
- **Local-first** — Designed for local development and sovereign VPS deployment
- **No phone-home** — All communication is user-controlled
- **Open source** — Full codebase on Gitea
- **Fleet-integrated** — Connects to sovereign agent fleet via MemPalace tunnels
- **Containerized** — Docker support for isolated deployment
**Verdict: Fully sovereign. 3D visualization + MUD + memory system in one integrated platform.**
---
*"It is meant to become two things at once: a local-first training ground for Timmy and a wizardly visualization surface for the living system."*

View File

@@ -0,0 +1,159 @@
# GENOME.md: turboquant
**Generated:** 2026-04-14
**Repo:** Timmy_Foundation/turboquant
**Phase:** 1 Complete (PolarQuant MVP)
**Status:** Production-ready Metal shaders, benchmarks passing
---
## Project Overview
TurboQuant is a KV cache compression system for local LLM inference on Apple Silicon. It enables 64K-128K context windows on 27B-parameter models within 36GB unified memory.
Three-stage compression pipeline:
1. **PolarQuant** -- WHT rotation + polar coordinates + Lloyd-Max codebook (~4.2x compression)
2. **QJL** -- 1-bit quantized Johnson-Lindenstrauss residual correction
3. **TurboQuant** -- PolarQuant + QJL combined = ~3.5 bits/channel, zero accuracy loss
**Key result:** turbo4 delivers 73% KV memory savings with 1% prompt processing overhead and 11% generation overhead.
## Architecture
```mermaid
graph TD
A[LLM Inference] --> B[KV Cache Layer]
B --> C{TurboQuant Mode}
C -->|turbo2| D[2-bit PolarQuant]
C -->|turbo3| E[3-bit PolarQuant + QJL]
C -->|turbo4| F[4-bit PolarQuant]
D --> G[Metal Shader: encode/decode]
E --> G
F --> G
G --> H[Compressed KV Storage]
H --> I[Decompress on Attention]
I --> A
J[llama-turbo.h] --> K[polar_quant_encode_turbo4]
K --> L[WHT Rotation]
L --> M[Polar Transform]
M --> N[Lloyd-Max Quantize]
N --> O[Packed 4-bit Output]
```
## Entry Points
| Entry Point | Type | Purpose |
|-------------|------|---------|
| `llama-turbo.cpp` | C++ library | Core encode/decode functions |
| `llama-turbo.h` | C header | Public API: `polar_quant_encode_turbo4`, `polar_quant_decode_turbo4` |
| `ggml-metal-turbo.metal` | Metal shader | GPU-accelerated encode/decode for Apple Silicon |
| `benchmarks/run_benchmarks.py` | Python | Benchmark suite: perplexity, speed, memory |
| `benchmarks/run_perplexity.py` | Python | Perplexity evaluation across context lengths |
| `evolution/hardware_optimizer.py` | Python | Hardware-aware parameter tuning |
| `.gitea/workflows/smoke.yml` | CI | Smoke test on push |
## Data Flow
```
Input: float array [d] (d=128, one KV head)
|
v
WHT Rotation (structured orthogonal transform)
|
v
Polar Transform (cartesian -> polar coordinates)
|
v
Lloyd-Max Quantization (non-uniform codebook, 4-bit)
|
v
Output: packed uint8_t [d/2] + float norm (radius)
```
Decode is the inverse: unpack -> dequantize -> inverse polar -> inverse WHT.
## Key Abstractions
| Abstraction | Description |
|-------------|-------------|
| **Turbo4** | 4-bit PolarQuant mode. Best quality. 4.2x compression. |
| **Turbo3** | 3-bit mode with QJL residual. ~3.5 bits/channel. |
| **Turbo2** | 2-bit mode. Maximum compression. Quality tradeoff. |
| **WHT** | Walsh-Hadamard Transform. Structured orthogonal rotation. |
| **Lloyd-Max** | Non-uniform codebook optimized for N(0, 1/sqrt(128)) distribution. |
| **QJL** | Quantized Johnson-Lindenstrauss. 1-bit residual correction. |
## API Surface
### C API (llama-turbo.h)
```c
// Encode: float [d] -> packed 4-bit [d/2] + norm
void polar_quant_encode_turbo4(const float* src, uint8_t* dst, float* norm, int d);
// Decode: packed 4-bit [d/2] + norm -> float [d]
void polar_quant_decode_turbo4(const uint8_t* src, float* dst, float norm, int d);
```
### Integration
TurboQuant integrates with llama.cpp via:
- Mixed quantization pairs: `q8_0 x turbo` for K/V asymmetric compression
- Metal shader dispatch in `ggml-metal.metal` (turbo kernels)
- Build flag: `-DGGML_TURBOQUANT=ON`
### Hermes Profile
`profiles/hermes-profile-gemma4-turboquant.yaml` defines deployment config:
- Model: gemma4 with turbo4 KV compression
- Target hardware: M3/M4 Max, 36GB+
- Context window: up to 128K with compression
## Test Coverage
| Area | Coverage | Notes |
|------|----------|-------|
| WHT rotation | Partial | Metal GPU uses WHT. CPU ref uses dense random (legacy). |
| Encode/decode symmetry | Full | `turbo_rotate_forward()` == inverse of `turbo_rotate_inverse()` |
| Lloyd-Max codebook | Full | Non-uniform centroids verified |
| Radius precision | Full | FP16+ norm per 128-element block |
| Metal shader correctness | Full | All dk32-dk576 variants tested |
| Perplexity benchmarks | Full | WikiText results in `benchmarks/perplexity_results.json` |
### Gaps
- No CI integration for Metal shader tests (smoke test only covers build)
- CPU reference implementation uses dense random, not WHT (legacy)
- No long-session stress tests beyond 128K
- QJL implementation not yet verified against CUDA reference
## Security Considerations
- **No network access.** All inference is local.
- **No user data in repo.** Benchmarks use public WikiText corpus.
- **Binary blobs.** `llama-turbo.cpp` compiles to native code. No sandboxing.
- **Upstream dependency.** Fork of TheTom/llama-cpp-turboquant. Trust boundary at upstream.
## Dependencies
| Dependency | Type | Source |
|------------|------|--------|
| llama.cpp | Fork | TheTom/llama-cpp-turboquant |
| Metal | System | Apple GPU framework |
| CMake | Build | Standard build system |
| Python 3.10+ | Scripts | Benchmarks and optimizer |
## Key Files
```
turboquant/
llama-turbo.h # C API header
llama-turbo.cpp # Core encode/decode implementation
ggml-metal-turbo.metal # Metal GPU shaders
benchmarks/ # Perplexity and speed benchmarks
evolution/ # Hardware optimizer
profiles/ # Hermes deployment profile
docs/ # Project status and build spec
.gitea/workflows/ # CI smoke test
```

View File

@@ -0,0 +1,150 @@
# GENOME.md — TurboQuant (Timmy_Foundation/turboquant)
> Codebase Genome v1.1 | Refreshed 2026-04-18 | Repo 12/16 | Ref: #679
## Project Overview
**TurboQuant** is a KV cache compression system for local inference on Apple Silicon. Implements Google's ICLR 2026 paper to unlock 64K-128K context on 27B models within 32GB unified memory.
**Three-stage compression:**
1. **PolarQuant** — WHT rotation + polar coordinates + Lloyd-Max codebook (~4.2x compression)
2. **QJL** — 1-bit quantized Johnson-Lindenstrauss residual correction
3. **TurboQuant** — PolarQuant + QJL = ~3.5 bits/channel, zero accuracy loss
**Key result:** 73% KV memory savings with 1% prompt processing overhead, 11% generation overhead.
## Architecture
```mermaid
graph TD
subgraph "Compression Pipeline"
KV[Raw KV Cache fp16] --> WHT[WHT Rotation]
WHT --> POLAR[PolarQuant 4-bit]
POLAR --> QJL[QJL Residual]
QJL --> PACKED[Packed KV ~3.5bit]
end
subgraph "Metal Shaders"
PACKED --> DECODE[Polar Decode Kernel]
DECODE --> ATTEN[Flash Attention]
ATTEN --> OUTPUT[Model Output]
end
subgraph "Build System"
CMAKE[CMakeLists.txt] --> LIB[turboquant.a]
LIB --> TEST[turboquant_roundtrip_test]
LIB --> LLAMA[llama.cpp fork integration]
end
subgraph "Python Layer"
SELECTOR[quant_selector.py] --> MODELS[model_registry/]
MODELS --> PROFILE[hardware_profiles.py]
PROFILE --> DECISION[quantization decision]
end
```
## Entry Points
| Entry Point | File | Purpose |
|-------------|------|---------|
| `polar_quant_encode_turbo4()` | llama-turbo.cpp | Encode float KV → 4-bit packed |
| `polar_quant_decode_turbo4()` | llama-turbo.cpp | Decode 4-bit packed → float KV |
| `cmake -S . -B build -DTURBOQUANT_BUILD_TESTS=ON` | CMakeLists.txt | Build static library + CTest suite |
| `ctest --test-dir build --output-on-failure` | build/ | Run C++ roundtrip tests |
| `run_benchmarks.py` | benchmarks/ | Run perplexity benchmarks |
| `quant_selector.py` | quant_selector/ | Hardware-aware quantization selection |
## Key Abstractions
| Symbol | File | Purpose |
|--------|------|---------|
| `polar_quant_encode_turbo4()` | llama-turbo.h/.cpp | Encode float[d] → packed 4-bit + L2 norm |
| `polar_quant_decode_turbo4()` | llama-turbo.h/.cpp | Decode packed 4-bit + norm → float[d] |
| `turbo_dequantize_k()` | ggml-metal-turbo.metal | Metal kernel: dequantize K cache |
| `turbo_dequantize_v()` | ggml-metal-turbo.metal | Metal kernel: dequantize V cache |
| `turbo_fwht_128()` | ggml-metal-turbo.metal | Fast Walsh-Hadamard Transform |
| `run_perplexity.py` | benchmarks/ | Measure perplexity impact |
| `run_benchmarks.py` | benchmarks/ | Full benchmark suite (speed + quality) |
| `select_quantization()` | quant_selector.py | Pick quant scheme from hardware profile |
## Data Flow
```
Input: float KV vectors [d=128 per head]
1. WHT rotation (in-place, O(d log d))
2. Convert to polar coords (radius, angles)
3. Lloyd-Max quantize angles → 4-bit indices
4. Store: packed indices [d/2 bytes] + float norm [4 bytes]
Decode: indices → codebook lookup → polar → cartesian → inverse WHT
Output: reconstructed float KV [d=128]
```
## API Surface
| Function | Signature | Notes |
|----------|-----------|-------|
| `polar_quant_encode_turbo4` | `(const float*, uint8_t*, float*, int)` | Core encode path |
| `polar_quant_decode_turbo4` | `(const uint8_t*, float, float*, int)` | Core decode path |
| `select_quantization` | `(HardwareProfile) -> QuantConfig` | Python quant selector |
## File Index
| File | LOC | Purpose |
|------|-----|---------|
| `llama-turbo.h` | 24 | C API: encode/decode function declarations |
| `llama-turbo.cpp` | 78 | Implementation: PolarQuant encode/decode |
| `ggml-metal-turbo.metal` | 76 | Metal shader: dequantize + FWHT kernels |
| `CMakeLists.txt` | 42 | Standalone build: lib + test targets |
| `quant_selector.py` | ~120 | Python: hardware profile → quant decision |
| `tests/test_quant_selector.py` | ~90 | Pytest: quant selector (currently failing) |
| `benchmarks/run_benchmarks.py` | ~85 | Perplexity + speed benchmarking |
## CI / Runtime Drift
| Dimension | Status | Notes |
|-----------|--------|-------|
| **CMake/CTest standalone build** | ✅ Passing | `cmake -S . -B build -DTURBOQUANT_BUILD_TESTS=ON && ctest --test-dir build` works on current main |
| **Python quant selector tests** | ❌ Failing | `tests/test_quant_selector.py` fails on current main — tracked in `turboquant #139` |
| **CI lane: quant_selector** | ❌ Broken | The quant selector CI lane is non-blocking due to persistent failures |
| **CI lane: cmake roundtrip** | ✅ Green | C++ roundtrip test passes in CI |
| **Metal shader compilation** | ⚠️ Apple Silicon only | Cannot be tested in CI runners; validated manually on M-series hardware |
## Test Coverage Gaps
- `tests/test_quant_selector.py` is currently broken — selector returns wrong quantization for edge-case hardware profiles (see `turboquant #139`)
- No CI coverage for Metal shader correctness (Apple Silicon only)
- Benchmark regression detection is manual; no automated threshold enforcement
## Security Considerations
- C API operates on caller-allocated buffers — no internal bounds checking on `d` parameter
- Python quant selector reads hardware profile from filesystem; path traversal risk if profile dir is user-controllable
## Dependencies
| Dependency | Version | Purpose |
|------------|---------|---------|
| CMake | ≥3.20 | Build system |
| Python | ≥3.10 | Benchmarks + quant selector |
| pytest | any | Test runner for Python tests |
| Metal (macOS) | 14+ | GPU shader compilation |
| llama.cpp | fork | Integration layer |
## Deployment
- Static library `turboquant.a` linked into llama.cpp fork
- Python quant selector invoked at model-load time to pick compression scheme
- No standalone server component; embedded in inference runtime
## Technical Debt
- `turboquant #139` — quant selector test failures not yet resolved; CI lane is non-blocking
- No automated benchmark regression detection
- Metal shaders untestable in CI — manual validation on Apple Silicon required
- Stale genome (v1.0, 2026-04-15) did not reflect quant selector addition or CI drift

263
genomes/wolf/GENOME.md Normal file
View File

@@ -0,0 +1,263 @@
# GENOME.md — Wolf (Timmy_Foundation/wolf)
> Codebase Genome v1.0 | Generated 2026-04-14 | Repo 16/16
## Project Overview
**Wolf** is a multi-model evaluation engine for sovereign AI fleets. It runs prompts against multiple LLM providers, scores responses on relevance, coherence, and safety, and outputs structured JSON results for model selection and ranking.
**Core principle:** agents work, PRs prove it, CI judges it.
**Status:** v1.0.0 — production-ready for prompt evaluation. Legacy PR evaluation module retained for backward compatibility.
## Architecture
```mermaid
graph TD
CLI[cli.py] --> Config[config.py]
CLI --> TaskGen[task.py]
CLI --> Runner[runner.py]
CLI --> Evaluator[evaluator.py]
CLI --> Leaderboard[leaderboard.py]
CLI --> Gitea[gitea.py]
Runner --> Models[models.py]
Runner --> Gitea
Evaluator --> Models
TaskGen --> Gitea
Leaderboard --> |leaderboard.json| FS[(File System)]
Config --> |wolf-config.yaml| FS
Models --> OpenRouter[OpenRouter API]
Models --> Groq[Groq API]
Models --> Ollama[Ollama Local]
Models --> OpenAI[OpenAI API]
Models --> Anthropic[Anthropic API]
Runner --> |branch + commit| Gitea
Evaluator --> |score results| Leaderboard
```
## Entry Points
| Entry Point | Command | Purpose |
|-------------|---------|---------|
| `wolf/cli.py` | `python3 -m wolf.cli --run` | Main CLI: run tasks, evaluate PRs, show leaderboard |
| `wolf/runner.py` | `python3 -m wolf.runner --prompts p.json --models m.json` | Standalone prompt evaluation runner |
| `wolf/__init__.py` | `import wolf` | Package init, version metadata |
## Data Flow
### Prompt Evaluation Pipeline (Primary)
```
prompts.json + models.json (or wolf-config.yaml)
PromptEvaluator.evaluate()
├─ For each (prompt, model) pair:
│ ├─ ModelClient.generate(prompt) → response text
│ ├─ ResponseScorer.score(response, prompt)
│ │ ├─ score_relevance() (0.40 weight)
│ │ ├─ score_coherence() (0.35 weight)
│ │ └─ score_safety() (0.25 weight)
│ └─ EvaluationResult (prompt, model, scores, latency, error)
evaluate_and_serialize() → JSON output
├─ model_summaries (per-model averages)
└─ results[] (per-evaluation details)
```
### Task Assignment Pipeline (Legacy)
```
Gitea Issues → TaskGenerator → AgentRunner
│ │ │
▼ ▼ ▼
Fetch tasks Assign models Execute + PR
from issues from config via Gitea API
```
## Key Abstractions
| Class | Module | Purpose |
|-------|--------|---------|
| `PromptEntry` | evaluator.py | Single prompt with expected keywords and category |
| `ModelEndpoint` | evaluator.py | Model connection descriptor (provider, model_id, key) |
| `ScoreResult` | evaluator.py | Scores for relevance, coherence, safety, overall |
| `EvaluationResult` | evaluator.py | Full result: prompt + model + response + scores + latency |
| `ResponseScorer` | evaluator.py | Heuristic scoring engine (regex + keyword + structure) |
| `PromptEvaluator` | evaluator.py | Core engine: runs prompts against models, scores output |
| `ModelClient` | models.py | Abstract base for LLM API calls |
| `ModelFactory` | models.py | Factory: returns correct client for provider name |
| `Task` | task.py | Work unit: id, title, description, assigned model/provider |
| `TaskGenerator` | task.py | Creates tasks from Gitea issues or JSON spec |
| `AgentRunner` | runner.py | Executes tasks: generate → branch → commit → PR |
| `Config` | config.py | YAML config loader (wolf-config.yaml) |
| `Leaderboard` | leaderboard.py | Persistent model ranking with serverless readiness |
| `GiteaClient` | gitea.py | Full Gitea REST API client |
| `PREvaluator` | evaluator.py | Legacy: scores PRs on CI, commits, code quality |
## API Surface
### CLI Arguments (cli.py)
| Flag | Description |
|------|-------------|
| `--config` | Path to wolf-config.yaml |
| `--task-spec` | Path to task specification JSON |
| `--run` | Run pending tasks (assign models, execute, create PRs) |
| `--evaluate` | Evaluate open PRs and score them |
| `--leaderboard` | Show model rankings |
### CLI Arguments (runner.py)
| Flag | Description |
|------|-------------|
| `--prompts` / `-p` | Path to prompts JSON (required) |
| `--models` / `-m` | Path to models JSON |
| `--config` / `-c` | Path to wolf-config.yaml (alternative to --models) |
| `--output` / `-o` | Path to write JSON results |
| `--system-prompt` | System prompt for all model calls |
### Provider Clients (models.py)
| Client | Provider | API Format |
|--------|----------|------------|
| `OpenRouterClient` | openrouter | OpenAI-compatible chat completions |
| `GroqClient` | groq | OpenAI-compatible chat completions |
| `OllamaClient` | ollama | Ollama native /api/generate |
| `OpenAIClient` | openai | OpenAI-compatible (reuses GroqClient with different URL) |
| `AnthropicClient` | anthropic | Anthropic Messages API v1 |
### Gitea Client (gitea.py)
| Method | Purpose |
|--------|---------|
| `get_issues()` | Fetch issues by state |
| `create_branch()` | Create new branch from base |
| `create_file()` | Create file on branch (base64) |
| `update_file()` | Update file with SHA |
| `get_file()` | Read file contents |
| `create_pull_request()` | Open PR |
| `get_pull_request()` | Fetch PR details |
| `get_pr_status()` | Check PR CI status |
## Configuration (wolf-config.yaml)
```yaml
gitea:
base_url: "https://forge.alexanderwhitestone.com/api/v1"
token: "..."
owner: "Timmy_Foundation"
repo: "wolf"
providers:
openrouter:
api_key: "..."
base_url: "https://openrouter.ai/api/v1"
ollama:
base_url: "http://localhost:11434"
models:
- model: "anthropic/claude-3.5-sonnet"
provider: "openrouter"
- model: "gemma4:latest"
provider: "ollama"
log_dir: "~/.hermes/wolf/"
leaderboard_path: "~/.hermes/wolf/leaderboard.json"
```
## Scoring Methodology
| Dimension | Weight | Method |
|-----------|--------|--------|
| Relevance | 0.40 | Keyword matching (expected_keywords), prompt word overlap, length penalty, refusal detection |
| Coherence | 0.35 | Structure indicators (paragraphs, lists, transitions), sentence completeness, repetition ratio, length sweet spot |
| Safety | 0.25 | Unsafe pattern regex (bomb/hack/bypass/injection), profanity detection |
**Overall = relevance×0.40 + coherence×0.35 + safety×0.25**
### Safety Patterns Detected
- Bomb/explosive/weapon creation
- System/network hacking
- Security/auth bypass
- Prompt injection ("ignore previous instructions")
- System prompt extraction attempts
### Refusal Patterns Detected
- "I cannot/can't/won't help/assist"
- "Sorry, but I cannot"
- "Against my guidelines/policy"
## Test Coverage
| File | Tests | Coverage |
|------|-------|----------|
| `tests/test_evaluator.py` | 17 tests | PromptEntry, ModelEndpoint, ResponseScorer (relevance/coherence/safety), PromptEvaluator (evaluate, error handling, serialization, file output, multi-model), PREvaluator (score_pr, description scoring) |
| `tests/test_config.py` | 1 test | Config load from YAML |
### Coverage Gaps
- No tests for `cli.py` (argument parsing, workflow orchestration)
- No tests for `runner.py` (`load_prompts`, `load_models_from_json`, `AgentRunner.execute_task`)
- No tests for `task.py` (`TaskGenerator.from_gitea_issues`, `from_spec`, `assign_tasks`)
- No tests for `models.py` (API clients — would require mocking HTTP)
- No tests for `leaderboard.py` (`record_score`, `get_rankings`, serverless readiness logic)
- No tests for `gitea.py` (API client — would require mocking HTTP)
- No integration tests (end-to-end evaluation pipeline)
## Dependencies
| Dependency | Used By | Purpose |
|------------|---------|---------|
| `requests` | models.py, gitea.py | HTTP client for all API calls |
| `pyyaml` (optional) | config.py | YAML config parsing (falls back to line parser) |
## Security Considerations
1. **API keys in config**: wolf-config.yaml stores provider API keys in plaintext. File should be chmod 600 and excluded from git (already in .gitignore pattern via ~/.hermes/).
2. **Gitea token**: Full access token used for branch creation, file commits, and PR creation. Scoped access recommended.
3. **No input sanitization**: Prompts from Gitea issues are passed directly to models without filtering. Prompt injection risk for automated workflows.
4. **No rate limiting**: Model API calls are sequential with no backoff or rate limiting. Could exhaust API quotas.
5. **Legacy code reference**: `evaluator.py` references `Evaluator = PREvaluator` alias but `cli.py` imports `Evaluator` expecting the legacy class. This works but is confusing.
## File Index
| File | LOC | Purpose |
|------|-----|---------|
| `wolf/__init__.py` | 12 | Package init, version |
| `wolf/cli.py` | 90 | Main CLI orchestrator |
| `wolf/config.py` | 48 | YAML config loader |
| `wolf/models.py` | 130 | LLM provider clients (5 providers) |
| `wolf/runner.py` | 280 | Prompt evaluation CLI + AgentRunner |
| `wolf/task.py` | 80 | Task dataclass + generator |
| `wolf/evaluator.py` | 350 | Core scoring engine + legacy PR evaluator |
| `wolf/leaderboard.py` | 70 | Persistent model ranking |
| `wolf/gitea.py` | 100 | Gitea REST API client |
| `tests/test_evaluator.py` | 180 | Unit tests for evaluator |
| `tests/test_config.py` | 20 | Unit tests for config |
**Total: ~1,360 LOC Python | 11 modules | 18 tests**
## Sovereignty Assessment
- **No external dependencies beyond requests**: Runs on any machine with Python 3.11+ and requests.
- **No phone-home**: All API calls are to user-configured endpoints.
- **No telemetry**: Logs go to local filesystem only.
- **Config-driven**: All secrets in user's ~/.hermes/ directory.
- **Provider-agnostic**: Supports 5 providers with easy extension via ModelFactory.
**Verdict: Fully sovereign. No corporate lock-in. User controls all endpoints and keys.**
---
*"The strength of the pack is the wolf, and the strength of the wolf is the pack."*
*— The Wolf Sovereign Core has spoken.*

View File

@@ -0,0 +1,142 @@
# Emacs Sovereign Control Plane
Real-time, programmable orchestration hub for the Timmy Foundation fleet.
## Architecture
```
┌─────────────────────────────────────────────────────────────┐
│ Emacs Control Plane │
│ │
│ ┌─────────────┐ ┌─────────────┐ ┌─────────────┐ │
│ │ dispatch.org│ │ shared │ │ org-babel │ │
│ │ (Task Queue)│ │ buffers │ │ notebooks │ │
│ └──────┬──────┘ └──────┬──────┘ └──────┬──────┘ │
│ │ │ │ │
│ └────────────────┼────────────────┘ │
│ │ │
│ ┌─────▼─────┐ │
│ │ Emacs │ │
│ │ Daemon │ │
│ │ (bezalel)│ │
│ └─────┬─────┘ │
└──────────────────────────┼──────────────────────────────────┘
┌──────────────────┼──────────────────┐
│ │ │
┌────▼────┐ ┌────▼────┐ ┌────▼────┐
│ Ezra │ │ Allegro │ │ Timmy │
│ (VPS) │ │ (VPS) │ │ (Mac) │
└─────────┘ └─────────┘ └─────────┘
```
## Infrastructure
| Component | Location | Details |
|-----------|----------|---------|
| Daemon Host | Bezalel (`159.203.146.185`) | Shared Emacs daemon |
| Socket Path | `/root/.emacs.d/server/bezalel` | emacsclient socket |
| Dispatch Hub | `/srv/fleet/workspace/dispatch.org` | Central task queue |
| Wrapper | `/usr/local/bin/fleet-append` | Quick message append |
## Quick Start
### From Local Machine (Timmy)
```bash
# Append a message to the fleet log
scripts/fleet_dispatch.sh append "Status: all systems nominal"
# Check for pending tasks assigned to Timmy
scripts/fleet_dispatch.sh poll timmy
# Claim a task
scripts/fleet_dispatch.sh claim 42 timmy
# Report task completion
scripts/emacs_fleet_bridge.py complete 42 "PR merged: #123"
```
### From Other VPS Agents (Ezra, Allegro, etc.)
```bash
# Direct emacsclient via SSH
ssh root@bezalel 'emacsclient -s /root/.emacs.d/server/bezalel -e "(your-elisp-here)"'
# Or use the wrapper
ssh root@bezalel '/usr/local/bin/fleet-append "Ezra: task #42 complete"'
```
## dispatch.org Structure
The central dispatch hub uses Org mode format:
```org
* TODO [timmy] Review PR #123 from gitea
SCHEDULED: <2026-04-13 Sun>
:PROPERTIES:
:PRIORITY: A
:ASSIGNEE: timmy
:GITEA_PR: https://forge.alexanderwhitestone.com/...
:END:
* IN_PROGRESS [ezra] Deploy monitoring to VPS
SCHEDULED: <2026-04-13 Sun>
:PROPERTIES:
:PRIORITY: B
:ASSIGNEE: ezra
:STARTED: 2026-04-13T15:30:00Z
:END:
* DONE [allegro] Fix cron reliability
CLOSED: [2026-04-13 Sun 14:00]
:PROPERTIES:
:ASSIGNEE: allegro
:RESULT: PR #456 merged
:END:
```
### Status Keywords
- `TODO` — Available for claiming
- `IN_PROGRESS` — Being worked on
- `WAITING` — Blocked on external dependency
- `DONE` — Completed
- `CANCELLED` — No longer needed
### Priority Levels
- `[#A]` — Critical / P0
- `[#B]` — Important / P1
- `[#C]` — Normal / P2
## Agent Workflow
1. **Poll:** Check `dispatch.org` for `TODO` items matching your agent name
2. **Claim:** Update status from `TODO` to `IN_PROGRESS`, add `:STARTED:` timestamp
3. **Execute:** Do the work (implement, deploy, test, etc.)
4. **Report:** Update status to `DONE`, add `:RESULT:` property with outcome
## Integration with Existing Systems
### Gitea Issues
- `dispatch.org` tasks can reference Gitea issues via `:GITEA_PR:` or `:GITEA_ISSUE:` properties
- Completion can auto-close Gitea issues via API
### Hermes Cron
- Hermes cron jobs can check `dispatch.org` before running
- Tasks in `dispatch.org` take priority over ambient issue burning
### Nostr Protocol
- Heartbeats still go through Nostr (kind 1)
- `dispatch.org` is for tactical coordination, Nostr is for strategic announcements
## Files
```
infrastructure/emacs-control-plane/
├── README.md # This file
├── dispatch.org.template # Template dispatch file
└── fleet_bridge.el # Emacs Lisp helpers
scripts/
├── fleet_dispatch.sh # Shell wrapper for fleet operations
├── emacs_fleet_bridge.py # Python bridge for Emacs daemon
└── emacs_task_poller.py # Poll for tasks assigned to an agent
```

View File

@@ -0,0 +1,50 @@
#+TITLE: Fleet Dispatch Hub
#+AUTHOR: Timmy Foundation
#+DATE: 2026-04-13
#+PROPERTY: header-args :tangle no
* Overview
This is the central task queue for the Timmy Foundation fleet.
Agents poll this file for =TODO= items matching their name.
* How to Use
1. Agents: Poll for =TODO= items with your assignee tag
2. Claim: Move to =IN_PROGRESS= with =:STARTED:= timestamp
3. Complete: Move to =DONE= with =:RESULT:= property
* Fleet Status
** Heartbeats
- timmy: LAST_HEARTBEAT <2026-04-13 Sun 15:00>
- ezra: LAST_HEARTBEAT <2026-04-13 Sun 15:00>
- allegro: LAST_HEARTBEAT <2026-04-13 Sun 14:55>
- bezalel: LAST_HEARTBEAT <2026-04-13 Sun 15:00>
* Tasks
** TODO [timmy] Example task — review pending PRs
SCHEDULED: <2026-04-13 Sun>
:PROPERTIES:
:PRIORITY: B
:ASSIGNEE: timmy
:GITEA_ISSUE: https://forge.alexanderwhitestone.com/Timmy_Foundation/timmy-home/issues/590
:END:
Check all open PRs across fleet repos and triage.
** TODO [ezra] Example task — run fleet health check
SCHEDULED: <2026-04-13 Sun>
:PROPERTIES:
:PRIORITY: C
:ASSIGNEE: ezra
:END:
SSH into each VPS and verify services are running.
** TODO [allegro] Example task — update cron job configs
SCHEDULED: <2026-04-13 Sun>
:PROPERTIES:
:PRIORITY: C
:ASSIGNEE: allegro
:END:
Review and update cron job definitions in timmy-config.
* Completed
#+BEGIN: clocktable :scope file :maxlevel 2
#+END:

1
pipelines/__init__.py Normal file
View File

@@ -0,0 +1 @@
"""Codebase genome pipeline helpers."""

View File

@@ -0,0 +1,6 @@
#!/usr/bin/env python3
from codebase_genome import main
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,557 @@
#!/usr/bin/env python3
"""Generate a deterministic GENOME.md for a repository."""
from __future__ import annotations
import argparse
import ast
import os
import re
from pathlib import Path
from typing import NamedTuple
IGNORED_DIRS = {
".git",
".hg",
".svn",
".venv",
"venv",
"node_modules",
"__pycache__",
".mypy_cache",
".pytest_cache",
"dist",
"build",
"coverage",
}
TEXT_SUFFIXES = {
".py",
".js",
".mjs",
".cjs",
".ts",
".tsx",
".jsx",
".html",
".css",
".md",
".txt",
".json",
".yaml",
".yml",
".sh",
".ini",
".cfg",
".toml",
}
SOURCE_SUFFIXES = {".py", ".js", ".mjs", ".cjs", ".ts", ".tsx", ".jsx", ".sh"}
DOC_FILENAMES = {"README.md", "CONTRIBUTING.md", "SOUL.md"}
class RepoFile(NamedTuple):
path: str
abs_path: Path
size_bytes: int
line_count: int
kind: str
class RunSummary(NamedTuple):
markdown: str
source_count: int
test_count: int
doc_count: int
def _is_text_file(path: Path) -> bool:
return path.suffix.lower() in TEXT_SUFFIXES or path.name in {"Dockerfile", "Makefile"}
def _file_kind(rel_path: str, path: Path) -> str:
suffix = path.suffix.lower()
if rel_path.startswith("tests/") or path.name.startswith("test_"):
return "test"
if rel_path.startswith("docs/") or path.name in DOC_FILENAMES or suffix == ".md":
return "doc"
if suffix in {".json", ".yaml", ".yml", ".toml", ".ini", ".cfg"}:
return "config"
if suffix == ".sh":
return "script"
if rel_path.startswith("scripts/") and suffix == ".py" and path.name != "__init__.py":
return "script"
if suffix in SOURCE_SUFFIXES:
return "source"
return "other"
def collect_repo_files(repo_root: str | Path) -> list[RepoFile]:
root = Path(repo_root).resolve()
files: list[RepoFile] = []
for current_root, dirnames, filenames in os.walk(root):
dirnames[:] = sorted(d for d in dirnames if d not in IGNORED_DIRS)
base = Path(current_root)
for filename in sorted(filenames):
path = base / filename
if not _is_text_file(path):
continue
rel_path = path.relative_to(root).as_posix()
text = path.read_text(encoding="utf-8", errors="replace")
files.append(
RepoFile(
path=rel_path,
abs_path=path,
size_bytes=path.stat().st_size,
line_count=max(1, len(text.splitlines())),
kind=_file_kind(rel_path, path),
)
)
return sorted(files, key=lambda item: item.path)
def _safe_text(path: Path) -> str:
return path.read_text(encoding="utf-8", errors="replace")
def _sanitize_node_id(name: str) -> str:
cleaned = re.sub(r"[^A-Za-z0-9_]", "_", name)
return cleaned or "node"
def _component_name(path: str) -> str:
if "/" in path:
return path.split("/", 1)[0]
return Path(path).stem or path
def _priority_files(files: list[RepoFile], kinds: tuple[str, ...], limit: int = 8) -> list[RepoFile]:
items = [item for item in files if item.kind in kinds]
items.sort(key=lambda item: (-int(item.path.count("/") == 0), -item.line_count, item.path))
return items[:limit]
def _readme_summary(root: Path) -> str:
readme = root / "README.md"
if not readme.exists():
return "Repository-specific overview missing from README.md. Genome generated from code structure and tests."
paragraphs: list[str] = []
current: list[str] = []
for raw_line in _safe_text(readme).splitlines():
line = raw_line.strip()
if not line:
if current:
paragraphs.append(" ".join(current).strip())
current = []
continue
if line.startswith("#"):
continue
current.append(line)
if current:
paragraphs.append(" ".join(current).strip())
return paragraphs[0] if paragraphs else "README.md exists but does not contain a prose overview paragraph."
def _extract_python_imports(text: str) -> set[str]:
try:
tree = ast.parse(text)
except SyntaxError:
return set()
imports: set[str] = set()
for node in ast.walk(tree):
if isinstance(node, ast.Import):
for alias in node.names:
imports.add(alias.name.split(".", 1)[0])
elif isinstance(node, ast.ImportFrom):
if node.module:
imports.add(node.module.split(".", 1)[0])
return imports
def _extract_python_symbols(text: str) -> tuple[list[tuple[str, int]], list[tuple[str, int]]]:
try:
tree = ast.parse(text)
except SyntaxError:
return [], []
classes: list[tuple[str, int]] = []
functions: list[tuple[str, int]] = []
for node in tree.body:
if isinstance(node, ast.ClassDef):
classes.append((node.name, node.lineno))
elif isinstance(node, ast.FunctionDef):
functions.append((node.name, node.lineno))
return classes, functions
def _build_component_edges(files: list[RepoFile]) -> list[tuple[str, str]]:
known_components = {_component_name(item.path) for item in files if item.kind in {"source", "script", "test"}}
edges: set[tuple[str, str]] = set()
for item in files:
if item.kind not in {"source", "script", "test"} or item.abs_path.suffix.lower() != ".py":
continue
src = _component_name(item.path)
imports = _extract_python_imports(_safe_text(item.abs_path))
for imported in imports:
if imported in known_components and imported != src:
edges.add((src, imported))
return sorted(edges)
def _render_mermaid(files: list[RepoFile]) -> str:
components = sorted(
{
_component_name(item.path)
for item in files
if item.kind in {"source", "script", "test", "config"}
and not _component_name(item.path).startswith(".")
}
)
edges = _build_component_edges(files)
lines = ["graph TD"]
if not components:
lines.append(" repo[\"repository\"]")
return "\n".join(lines)
for component in components[:10]:
node_id = _sanitize_node_id(component)
lines.append(f" {node_id}[\"{component}\"]")
seen_components = set(components[:10])
emitted = False
for src, dst in edges:
if src in seen_components and dst in seen_components:
lines.append(f" {_sanitize_node_id(src)} --> {_sanitize_node_id(dst)}")
emitted = True
if not emitted:
root_id = "repo_root"
lines.insert(1, f" {root_id}[\"repo\"]")
for component in components[:6]:
lines.append(f" {root_id} --> {_sanitize_node_id(component)}")
return "\n".join(lines)
def _entry_points(files: list[RepoFile]) -> list[dict[str, str]]:
points: list[dict[str, str]] = []
for item in files:
text = _safe_text(item.abs_path)
if item.kind == "script":
points.append({"path": item.path, "reason": "operational script", "command": f"python3 {item.path}" if item.abs_path.suffix == ".py" else f"bash {item.path}"})
continue
if item.abs_path.suffix == ".py" and "if __name__ == '__main__':" in text:
points.append({"path": item.path, "reason": "python main guard", "command": f"python3 {item.path}"})
elif item.path in {"app.py", "server.py", "main.py"}:
points.append({"path": item.path, "reason": "top-level executable", "command": f"python3 {item.path}"})
seen: set[str] = set()
deduped: list[dict[str, str]] = []
for point in points:
if point["path"] in seen:
continue
seen.add(point["path"])
deduped.append(point)
return deduped[:12]
def _test_coverage(files: list[RepoFile]) -> tuple[list[RepoFile], list[RepoFile], list[RepoFile]]:
source_files = [
item
for item in files
if item.kind in {"source", "script"}
and item.path not in {"pipelines/codebase-genome.py", "pipelines/codebase_genome.py"}
and not item.path.endswith("/__init__.py")
]
test_files = [item for item in files if item.kind == "test"]
combined_test_text = "\n".join(_safe_text(item.abs_path) for item in test_files)
entry_paths = {point["path"] for point in _entry_points(files)}
gaps: list[RepoFile] = []
for item in source_files:
stem = item.abs_path.stem
if item.path in entry_paths:
continue
if stem and stem in combined_test_text:
continue
gaps.append(item)
gaps.sort(key=lambda item: (-item.line_count, item.path))
return source_files, test_files, gaps
def _security_findings(files: list[RepoFile]) -> list[dict[str, str]]:
rules = [
("high", "shell execution", re.compile(r"shell\s*=\s*True"), "shell=True expands blast radius for command execution"),
("high", "dynamic evaluation", re.compile(r"\b(eval|exec)\s*\("), "dynamic evaluation bypasses static guarantees"),
("medium", "unsafe deserialization", re.compile(r"pickle\.load\(|yaml\.load\("), "deserialization of untrusted data can execute code"),
("medium", "network egress", re.compile(r"urllib\.request\.urlopen\(|requests\.(get|post|put|delete)\("), "outbound network calls create runtime dependency and failure surface"),
("medium", "hardcoded http endpoint", re.compile(r"http://[^\s\"']+"), "plaintext or fixed HTTP endpoints can drift or leak across environments"),
]
findings: list[dict[str, str]] = []
for item in files:
if item.kind not in {"source", "script", "config"}:
continue
for lineno, line in enumerate(_safe_text(item.abs_path).splitlines(), start=1):
for severity, category, pattern, detail in rules:
if pattern.search(line):
findings.append(
{
"severity": severity,
"category": category,
"ref": f"{item.path}:{lineno}",
"line": line.strip(),
"detail": detail,
}
)
break
if len(findings) >= 12:
return findings
return findings
def _dead_code_candidates(files: list[RepoFile]) -> list[RepoFile]:
source_files = [item for item in files if item.kind in {"source", "script"} and item.abs_path.suffix == ".py"]
imports_by_file = {
item.path: _extract_python_imports(_safe_text(item.abs_path))
for item in source_files
}
imported_names = {name for imports in imports_by_file.values() for name in imports}
referenced_by_tests = "\n".join(_safe_text(item.abs_path) for item in files if item.kind == "test")
entry_paths = {point["path"] for point in _entry_points(files)}
candidates: list[RepoFile] = []
for item in source_files:
stem = item.abs_path.stem
if item.path in entry_paths:
continue
if stem in imported_names:
continue
if stem in referenced_by_tests:
continue
if stem in {"__init__", "conftest"}:
continue
candidates.append(item)
candidates.sort(key=lambda item: (-item.line_count, item.path))
return candidates[:10]
def _performance_findings(files: list[RepoFile]) -> list[dict[str, str]]:
findings: list[dict[str, str]] = []
for item in files:
if item.kind in {"source", "script"} and item.line_count >= 350:
findings.append({
"ref": item.path,
"detail": f"large module ({item.line_count} lines) likely hides multiple responsibilities",
})
for item in files:
if item.kind not in {"source", "script"}:
continue
text = _safe_text(item.abs_path)
if "os.walk(" in text or ".rglob(" in text or "glob.glob(" in text:
findings.append({
"ref": item.path,
"detail": "per-run filesystem scan detected; performance scales with repo size",
})
if "urllib.request.urlopen(" in text or "requests.get(" in text or "requests.post(" in text:
findings.append({
"ref": item.path,
"detail": "network-bound execution path can dominate runtime and create flaky throughput",
})
deduped: list[dict[str, str]] = []
seen: set[tuple[str, str]] = set()
for finding in findings:
key = (finding["ref"], finding["detail"])
if key in seen:
continue
seen.add(key)
deduped.append(finding)
return deduped[:10]
def _key_abstractions(files: list[RepoFile]) -> list[dict[str, object]]:
abstractions: list[dict[str, object]] = []
for item in _priority_files(files, ("source", "script"), limit=10):
if item.abs_path.suffix != ".py":
continue
classes, functions = _extract_python_symbols(_safe_text(item.abs_path))
if not classes and not functions:
continue
abstractions.append(
{
"path": item.path,
"classes": classes[:4],
"functions": [entry for entry in functions[:6] if not entry[0].startswith("_")],
}
)
return abstractions[:8]
def _api_surface(entry_points: list[dict[str, str]], abstractions: list[dict[str, object]]) -> list[str]:
api_lines: list[str] = []
for entry in entry_points[:8]:
api_lines.append(f"- CLI: `{entry['command']}` — {entry['reason']} (`{entry['path']}`)")
for abstraction in abstractions[:5]:
for func_name, lineno in abstraction["functions"]:
api_lines.append(f"- Python: `{func_name}()` from `{abstraction['path']}:{lineno}`")
if len(api_lines) >= 14:
return api_lines
return api_lines
def _data_flow(entry_points: list[dict[str, str]], files: list[RepoFile], gaps: list[RepoFile]) -> list[str]:
components = sorted(
{
_component_name(item.path)
for item in files
if item.kind in {"source", "script", "test", "config"} and not _component_name(item.path).startswith(".")
}
)
lines = []
if entry_points:
lines.append(f"1. Operators enter through {', '.join(f'`{item['path']}`' for item in entry_points[:3])}.")
else:
lines.append("1. No explicit CLI/main guard entry point was detected; execution appears library- or doc-driven.")
if components:
lines.append(f"2. Core logic fans into top-level components: {', '.join(f'`{name}`' for name in components[:6])}.")
if gaps:
lines.append(f"3. Validation is incomplete around {', '.join(f'`{item.path}`' for item in gaps[:3])}, so changes there carry regression risk.")
else:
lines.append("3. Tests appear to reference the currently indexed source set, reducing blind spots in the hot path.")
lines.append("4. Final artifacts land as repository files, docs, or runtime side effects depending on the selected entry point.")
return lines
def generate_genome_markdown(repo_root: str | Path, repo_name: str | None = None) -> str:
root = Path(repo_root).resolve()
files = collect_repo_files(root)
repo_display = repo_name or root.name
summary = _readme_summary(root)
entry_points = _entry_points(files)
source_files, test_files, coverage_gaps = _test_coverage(files)
security = _security_findings(files)
dead_code = _dead_code_candidates(files)
performance = _performance_findings(files)
abstractions = _key_abstractions(files)
api_surface = _api_surface(entry_points, abstractions)
data_flow = _data_flow(entry_points, files, coverage_gaps)
mermaid = _render_mermaid(files)
lines: list[str] = [
f"# GENOME.md — {repo_display}",
"",
"Generated by `pipelines/codebase_genome.py`.",
"",
"## Project Overview",
"",
summary,
"",
f"- Text files indexed: {len(files)}",
f"- Source and script files: {len(source_files)}",
f"- Test files: {len(test_files)}",
f"- Documentation files: {len([item for item in files if item.kind == 'doc'])}",
"",
"## Architecture",
"",
"```mermaid",
mermaid,
"```",
"",
"## Entry Points",
"",
]
if entry_points:
for item in entry_points:
lines.append(f"- `{item['path']}` — {item['reason']} (`{item['command']}`)")
else:
lines.append("- No explicit entry point detected.")
lines.extend(["", "## Data Flow", ""])
lines.extend(data_flow)
lines.extend(["", "## Key Abstractions", ""])
if abstractions:
for abstraction in abstractions:
path = abstraction["path"]
classes = abstraction["classes"]
functions = abstraction["functions"]
class_bits = ", ".join(f"`{name}`:{lineno}" for name, lineno in classes) or "none detected"
function_bits = ", ".join(f"`{name}()`:{lineno}" for name, lineno in functions) or "none detected"
lines.append(f"- `{path}` — classes {class_bits}; functions {function_bits}")
else:
lines.append("- No Python classes or top-level functions detected in the highest-priority source files.")
lines.extend(["", "## API Surface", ""])
if api_surface:
lines.extend(api_surface)
else:
lines.append("- No obvious public API surface detected.")
lines.extend(["", "## Test Coverage Report", ""])
lines.append(f"- Source and script files inspected: {len(source_files)}")
lines.append(f"- Test files inspected: {len(test_files)}")
if coverage_gaps:
lines.append("- Coverage gaps:")
for item in coverage_gaps[:12]:
lines.append(f" - `{item.path}` — no matching test reference detected")
else:
lines.append("- No obvious coverage gaps detected by the stem-matching heuristic.")
lines.extend(["", "## Security Audit Findings", ""])
if security:
for finding in security:
lines.append(
f"- [{finding['severity']}] `{finding['ref']}` — {finding['category']}: {finding['detail']}. Evidence: `{finding['line']}`"
)
else:
lines.append("- No high-signal security findings detected by the static heuristics in this pass.")
lines.extend(["", "## Dead Code Candidates", ""])
if dead_code:
for item in dead_code:
lines.append(f"- `{item.path}` — not imported by indexed Python modules and not referenced by tests")
else:
lines.append("- No obvious dead-code candidates detected.")
lines.extend(["", "## Performance Bottleneck Analysis", ""])
if performance:
for finding in performance:
lines.append(f"- `{finding['ref']}` — {finding['detail']}")
else:
lines.append("- No obvious performance hotspots detected by the static heuristics in this pass.")
return "\n".join(lines).rstrip() + "\n"
def write_genome(repo_root: str | Path, repo_name: str | None = None, output_path: str | Path | None = None) -> RunSummary:
root = Path(repo_root).resolve()
markdown = generate_genome_markdown(root, repo_name=repo_name)
out_path = Path(output_path) if output_path else root / "GENOME.md"
out_path.parent.mkdir(parents=True, exist_ok=True)
out_path.write_text(markdown, encoding="utf-8")
files = collect_repo_files(root)
source_files, test_files, _ = _test_coverage(files)
return RunSummary(
markdown=markdown,
source_count=len(source_files),
test_count=len(test_files),
doc_count=len([item for item in files if item.kind == "doc"]),
)
def main() -> None:
parser = argparse.ArgumentParser(description="Generate a deterministic GENOME.md for a repository")
parser.add_argument("--repo-root", required=True, help="Path to the repository to analyze")
parser.add_argument("--repo", dest="repo_name", default=None, help="Optional repo display name")
parser.add_argument("--repo-name", dest="repo_name_override", default=None, help="Optional repo display name")
parser.add_argument("--output", default=None, help="Path to write GENOME.md (defaults to <repo-root>/GENOME.md)")
args = parser.parse_args()
repo_name = args.repo_name_override or args.repo_name
summary = write_genome(args.repo_root, repo_name=repo_name, output_path=args.output)
target = Path(args.output) if args.output else Path(args.repo_root).resolve() / "GENOME.md"
print(
f"GENOME.md saved to {target} "
f"(sources={summary.source_count}, tests={summary.test_count}, docs={summary.doc_count})"
)
if __name__ == "__main__":
main()

View File

@@ -1,124 +1,253 @@
# MemPalace Integration Evaluation Report
**Issue:** #568
**Original draft landed in:** PR #569
**Status:** Updated with live mining results, independent verification, and current recommendation
## Executive Summary
Evaluated **MemPalace v3.0.0** (github.com/milla-jovovich/mempalace) as a memory layer for the Timmy/Hermes agent stack.
Evaluated **MemPalace v3.0.0** (`github.com/milla-jovovich/mempalace`) as a memory layer for the Timmy/Hermes stack.
**Installed:**`mempalace 3.0.0` via `pip install`
**Works with:** ChromaDB, MCP servers, local LLMs
**Zero cloud:** ✅ Fully local, no API keys required
What is now established from the issue thread plus the merged draft:
- **Synthetic evaluation:** positive
- **Live mining on Timmy data:** positive
- **Independent Allegro verification:** positive
- **Zero-cloud property:** confirmed
- **Recommendation:** MemPalace is strong enough for pilot integration and wake-up experiments, but `timmy-home` should treat it as a proven candidate rather than the final uncontested winner until it is benchmarked against the current Engram direction documented elsewhere in this repo.
## Benchmark Findings (from Paper)
In other words: the evaluation succeeded. The remaining question is not whether MemPalace works. It is whether MemPalace should become the permanent fleet memory default.
## Benchmark Findings
These benchmark numbers were cited in the original evaluation draft:
| Benchmark | Mode | Score | API Required |
|---|---|---|---|
| **LongMemEval R@5** | Raw ChromaDB only | **96.6%** | **Zero** |
| **LongMemEval R@5** | Hybrid + Haiku rerank | **100%** | Optional Haiku |
| **LoCoMo R@10** | Raw, session level | 60.3% | Zero |
| **Personal palace R@10** | Heuristic bench | 85% | Zero |
| **Palace structure impact** | Wing+room filtering | **+34%** R@10 | Zero |
|---|---|---:|---|
| LongMemEval R@5 | Raw ChromaDB only | 96.6% | Zero |
| LongMemEval R@5 | Hybrid + Haiku rerank | 100% | Optional Haiku |
| LoCoMo R@10 | Raw, session level | 60.3% | Zero |
| Personal palace R@10 | Heuristic bench | 85% | Zero |
| Palace structure impact | Wing + room filtering | +34% R@10 | Zero |
## Before vs After Evaluation (Live Test)
These are paper-level or draft-level metrics. They matter, but the more important evidence for `timmy-home` is the live operational testing below.
### Test Setup
- Created test project with 4 files (README.md, auth.md, deployment.md, main.py)
- Mined into MemPalace palace
- Ran 4 standard queries
- Results recorded
## Before vs After Evaluation
### Before (Standard BM25 / Simple Search)
### Synthetic test setup
- 4-file test project:
- `README.md`
- `auth.md`
- `deployment.md`
- `main.py`
- mined into a MemPalace palace
- queried with 4 standard prompts
### Before (keyword/BM25 style expectations)
| Query | Would Return | Notes |
|---|---|---|
| "authentication" | auth.md (exact match only) | Misses context about JWT choice |
| "docker nginx SSL" | deployment.md | Manual regex/keyword matching needed |
| "keycloak OAuth" | auth.md | Would need full-text index |
| "postgresql database" | README.md (maybe) | Depends on index |
| `authentication` | `auth.md` | exact match only; weak on implementation context |
| `docker nginx SSL` | `deployment.md` | requires manual keyword logic |
| `keycloak OAuth` | `auth.md` | little semantic cross-reference |
| `postgresql database` | `README.md` maybe | depends on index quality |
**Problems:**
- No semantic understanding
- Exact match only
- No conversation memory
- No structured organization
- No wake-up context
Problems in the draft baseline:
- no semantic ranking
- exact match bias
- no durable conversation memory
- no palace structure
- no wake-up context artifact
### After (MemPalace)
### After (MemPalace synthetic results)
| Query | Results | Score | Notes |
|---|---|---:|---|
| `authentication` | `auth.md`, `main.py` | -0.139 | finds auth discussion and implementation |
| `docker nginx SSL` | `deployment.md`, `auth.md` | 0.447 | exact deployment hit plus related JWT context |
| `keycloak OAuth` | `auth.md`, `main.py` | -0.029 | finds both conceptual and implementation evidence |
| `postgresql database` | `README.md`, `main.py` | 0.025 | finds decision and implementation |
### Wake-up Context (synthetic)
- ~210 tokens total
- L0 identity placeholder
- L1 compressed project facts
- prompt-injection ready as a session wake-up payload
## Live Mining Results
Timmy later moved past the synthetic test and mined live agent context. That is the more important result for this repo.
### Live Timmy mining outcome
- **5,198 drawers** across 3 wings
- **413 files** mined from `~/.timmy/`
- wings reported in the issue:
- `timmy_soul` -> 27 drawers
- `timmy_memory` -> 5,166 drawers
- `mempalace-eval` -> 5 drawers
- **wake-up context:** ~785 tokens of L0 + L1
### Verified retrieval examples
Timmy reported successful verbatim retrieval for:
- `sovereignty service`
- exact SOUL.md text about sovereignty and service
- `crisis suicidal`
- exact crisis protocol text and related mission context
### Live before/after summary
| Query Type | Before MemPalace | After MemPalace | Delta |
|---|---|---|---|
| "authentication" | auth.md, main.py | -0.139 | Finds both auth discussion and JWT implementation |
| "docker nginx SSL" | deployment.md, auth.md | 0.447 | Exact match on deployment, related JWT context |
| "keycloak OAuth" | auth.md, main.py | -0.029 | Finds OAuth discussion and JWT usage |
| "postgresql database" | README.md, main.py | 0.025 | Finds both decision and implementation |
| Sovereignty facts | Model confabulation | Verbatim SOUL.md retrieval | 100% accuracy on the cited example |
| Crisis protocol | No persistent recall | Exact protocol text | Mission-critical recall restored |
| Config decisions | Lost between sessions | Persistent + searchable | Stops re-deciding known facts |
| Agent memory | Context window only | 5,198 searchable drawers | Large durable recall expansion |
| Wake-up tokens | 0 | ~785 compressed | Session-start context becomes possible |
### Wake-up Context
- **~210 tokens** total
- L0: Identity (placeholder)
- L1: All essential facts compressed
- Ready to inject into any LLM prompt
This is the strongest evidence in the issue: the evaluation moved from toy files to real Timmy memory material and still held up.
## Integration Potential
## Independent Verification
### 1. Memory Mining
Allegro independently reproduced the evaluation protocol.
### Allegro installation and setup
- installed `mempalace` in an isolated venv
- observed ChromaDB backend
- observed first-run embedding model download (~79MB)
- recreated the 4-file synthetic evaluation project
### Allegro before/after comparison
#### Before (grep/BM25 style)
| Query | Results | Limitation |
|---|---|---|
| `authentication` | 4 exact matches across 3 files | no ranking, no semantic context |
| `docker nginx SSL` | 7 matches across 5 files | OR-style matching, weak relevance ordering |
| `keycloak OAuth` | 5 matches across 4 files | exact keyword bias |
| `postgresql database` | 5 matches across 4 files | no cross-reference understanding |
#### After (MemPalace)
| Query | Top Result | Score | Cross-references Found |
|---|---|---:|---|
| `authentication` | `auth.md` | -0.232 | ranked related files behind it |
| `docker nginx SSL` | `main.py` | 0.073 | ranked infra-related files by relevance |
| `keycloak OAuth` | `auth.md` | -0.137 | related files surfaced as well |
| `postgresql database` | `README.md` | -0.084 | decision + implementation chain visible |
### Allegro wake-up context
- **238 tokens** generated
- identity placeholder still needed explicit setup
- confirmed there is a real compressed wake-up artifact, not just search hits
### Independent verdict
Allegro explicitly confirmed Timmy's core finding:
- semantic retrieval works
- palace structure is useful
- no cloud dependency is required
That matters because it reduces the chance that Timmy's result was a one-machine artifact.
## Operational Gotchas
The issue thread also surfaced practical constraints that matter more than the headline scores.
1. `mempalace init` is interactive even with `--yes`
- practical workaround: write `mempalace.yaml` manually
2. YAML schema gotcha
- key is `wing:` not `wings:`
- rooms are expected as a list of dicts
3. First-run download cost
- embedding model auto-download observed at ~79MB
- this is fine on a healthy machine but matters for cold-start and constrained hosts
4. Managed Python / venv dependency
- installation is straightforward, but it still assumes a controllable local Python environment
5. Integration is still only described, not fully landed
- the issue thread proposes:
- wake-up hook
- post-session mining
- MCP integration
- replacement of older memory paths
- those are recommendations and next steps, not completed mainline integration in `timmy-home`
## Recommendation
### Recommendation for this issue (#568)
**Accept the evaluation as successful and complete.**
MemPalace demonstrated:
- positive synthetic before/after improvement
- positive live Timmy mining results
- positive independent Allegro verification
- zero-cloud operation
- useful wake-up context generation
That is enough to say the evaluation question has been answered.
### Recommendation for `timmy-home` roadmap
**Do not overstate the result as “MemPalace is now the permanent uncontested memory layer.”**
A more precise current recommendation is:
1. use MemPalace as a proven pilot candidate for memory mining and wake-up experiments
2. keep the evaluation report as evidence that semantic local memory works in this stack
3. benchmark it against the current Engram direction before declaring final fleet-wide replacement
Why that caution is justified from inside this repo:
- `docs/hermes-agent-census.md` now treats **Engram memory provider** as a high-priority sovereignty path
- the issue thread proves MemPalace can work, but it does not prove MemPalace is the final best long-term provider for every host and workflow
### Practical call
- **For evaluation:** MemPalace passes
- **For immediate experimentation:** proceed
- **For irreversible architectural replacement:** compare against Engram first
## Integration Path Already Proposed
The issue thread and merged draft already outline a practical integration path worth preserving:
### Memory mining
```bash
# Mine Timmy's conversations
mempalace mine ~/.hermes/sessions/ --mode convos
# Mine project code and docs
mempalace mine ~/.hermes/hermes-agent/
# Mine configs
mempalace mine ~/.hermes/
```
### 2. Wake-up Protocol
### Wake-up protocol
```bash
mempalace wake-up > /tmp/timmy-context.txt
# Inject into Hermes system prompt
```
### 3. MCP Integration
### MCP integration
```bash
# Add as MCP tool
hermes mcp add mempalace -- python -m mempalace.mcp_server
```
### 4. Hermes Integration Pattern
- `PreCompact` hook: save memory before context compression
- `PostAPI` hook: mine conversation after significant interactions
- `WakeUp` hook: load context at session start
### Hook points suggested in the draft
- `PreCompact` hook
- `PostAPI` hook
- `WakeUp` hook
## Recommendations
These remain sensible as pilot integration points.
### Immediate
1. Add `mempalace` to Hermes venv requirements
2. Create mine script for ~/.hermes/ and ~/.timmy/
3. Add wake-up hook to Hermes session start
4. Test with real conversation exports
## Next Steps
### Short-term (Next Week)
1. Mine last 30 days of Timmy sessions
2. Build wake-up context for all agents
3. Add MemPalace MCP tools to Hermes toolset
4. Test retrieval quality on real queries
### Medium-term (Next Month)
1. Replace homebrew memory system with MemPalace
2. Build palace structure: wings for projects, halls for topics
3. Compress with AAAK for 30x storage efficiency
4. Benchmark against current RetainDB system
## Issues Filed
See Gitea issue #[NUMBER] for tracking.
Short list that follows directly from the evaluation without overcommitting the architecture:
- [ ] wire a MemPalace wake-up experiment into Hermes session start
- [ ] test post-session mining on real exported conversations
- [ ] measure retrieval quality on real operator queries, not only synthetic prompts
- [ ] run the same before/after protocol against Engram for a direct comparison
- [ ] only then decide whether MemPalace replaces or merely informs the permanent sovereign memory provider path
## Conclusion
MemPalace scores higher than published alternatives (Mem0, Mastra, Supermemory) with **zero API calls**.
PR #569 captured the first good draft of the MemPalace evaluation, but it left the issue open and the report unfinished.
For our use case, the key advantages are:
1. **Verbatim retrieval** — never loses the "why" context
2. **Palace structure** — +34% boost from organization
3. **Local-only** — aligns with our sovereignty mandate
4. **MCP compatible** — drops into our existing tool chain
5. **AAAK compression** — 30x storage reduction coming
This updated report closes the loop by consolidating:
- the original synthetic benchmarks
- Timmy's live mining results
- Allegro's independent verification
- the real operational gotchas
- a recommendation precise enough for the current `timmy-home` roadmap
It replaces the "we should build this" memory layer with something that already works and scores better than the research alternatives.
Bottom line:
- **MemPalace worked.**
- **The evaluation succeeded.**
- **The permanent memory-provider choice should still be made comparatively, not by enthusiasm alone.**

View File

@@ -0,0 +1,106 @@
# MemPalace v3.0.0 Integration — Before/After Evaluation
> Issue #568 | timmy-home
> Date: 2026-04-07
## Executive Summary
Evaluated **MemPalace v3.0.0** as a memory layer for the Timmy/Hermes agent stack.
**Installed:**`mempalace 3.0.0` via `pip install`
**Works with:** ChromaDB, MCP servers, local LLMs
**Zero cloud:** ✅ Fully local, no API keys required
## Benchmark Findings
| Benchmark | Mode | Score | API Required |
|-----------|------|-------|-------------|
| LongMemEval R@5 | Raw ChromaDB only | **96.6%** | **Zero** |
| LongMemEval R@5 | Hybrid + Haiku rerank | **100%** | Optional Haiku |
| LoCoMo R@10 | Raw, session level | 60.3% | Zero |
| Personal palace R@10 | Heuristic bench | 85% | Zero |
| Palace structure impact | Wing+room filtering | **+34%** R@10 | Zero |
## Before vs After (Live Test)
### Before (Standard BM25 / Simple Search)
- No semantic understanding
- Exact match only
- No conversation memory
- No structured organization
- No wake-up context
### After (MemPalace)
| Query | Results | Score | Notes |
|-------|---------|-------|-------|
| "authentication" | auth.md, main.py | -0.139 | Finds both auth discussion and JWT implementation |
| "docker nginx SSL" | deployment.md, auth.md | 0.447 | Exact match on deployment, related JWT context |
| "keycloak OAuth" | auth.md, main.py | -0.029 | Finds OAuth discussion and JWT usage |
| "postgresql database" | README.md, main.py | 0.025 | Finds both decision and implementation |
### Wake-up Context
- **~210 tokens** total
- L0: Identity (placeholder)
- L1: All essential facts compressed
- Ready to inject into any LLM prompt
## Integration Path
### 1. Memory Mining
```bash
mempalace mine ~/.hermes/sessions/ --mode convos
mempalace mine ~/.hermes/hermes-agent/
mempalace mine ~/.hermes/
```
### 2. Wake-up Protocol
```bash
mempalace wake-up > /tmp/timmy-context.txt
```
### 3. MCP Integration
```bash
hermes mcp add mempalace -- python -m mempalace.mcp_server
```
### 4. Hermes Hooks
- `PreCompact`: save memory before context compression
- `PostAPI`: mine conversation after significant interactions
- `WakeUp`: load context at session start
## Recommendations
### Immediate
1. Add `mempalace` to Hermes venv requirements
2. Create mine script for ~/.hermes/ and ~/.timmy/
3. Add wake-up hook to Hermes session start
4. Test with real conversation exports
### Short-term
1. Mine last 30 days of Timmy sessions
2. Build wake-up context for all agents
3. Add MemPalace MCP tools to Hermes toolset
4. Test retrieval quality on real queries
### Medium-term
1. Replace homebrew memory system with MemPalace
2. Build palace structure: wings for projects, halls for topics
3. Compress with AAAK for 30x storage efficiency
4. Benchmark against current RetainDB system
## Conclusion
MemPalace scores higher than published alternatives (Mem0, Mastra, Supermemory) with **zero API calls**.
Key advantages:
1. **Verbatim retrieval** — never loses the "why" context
2. **Palace structure** — +34% boost from organization
3. **Local-only** — aligns with sovereignty mandate
4. **MCP compatible** — drops into existing tool chain
5. **AAAK compression** — 30x storage reduction coming
---
*Evaluated by Timmy | Issue #568*

View File

@@ -0,0 +1,206 @@
# Phase 4 Sovereignty Audit
Generated: 2026-04-15 00:45:01 EDT
Issue: #551
Scope: repo-grounded audit of whether `timmy-home` currently proves **[PHASE-4] Sovereignty - Zero Cloud Dependencies**
## Phase Definition
Issue #551 defines Phase 4 as:
- no API call leaves your infrastructure
- no rate limits
- no censorship
- no shutdown dependency
- trigger condition: all Phase-3 buildings operational and all models running locally
The milestone sentence is explicit:
> “A model ran locally for the first time. No cloud. No rate limits. No one can turn it off.”
This audit asks a narrower, truthful question:
**Does the current `timmy-home` repo prove that the Timmy harness is already in Phase 4?**
## Current Repo Evidence
### 1. The repo already contains a local-only cutover diagnosis — and it says the harness is not there yet
Primary source:
- `specs/2026-03-29-local-only-harness-cutover-plan.md`
That plan records a live-state audit from 2026-03-29 and names concrete blockers:
- active cloud default in `~/.hermes/config.yaml`
- cloud fallback entries
- enabled cron inheritance risk
- legacy remote ops scripts still on the active path
- optional Groq offload still present in the Nexus path
Direct repo-grounded examples from that file:
- `model.default: gpt-5.4`
- `model.provider: openai-codex`
- `model.base_url: https://chatgpt.com/backend-api/codex`
- custom provider: Google Gemini
- fallback path still pointing to Gemini
- active cloud escape path via `groq_worker.py`
The same cutover plan defines “done” in stricter terms than the issue body and plainly says those conditions were not yet met.
### 2. The baseline report says sovereignty is still overwhelmingly cloud-backed
Primary source:
- `reports/production/2026-03-29-local-timmy-baseline.md`
That report gives the clearest quantitative evidence in this repo:
- sovereignty score: `0.7%` local
- sessions: `403 total | 3 local | 400 cloud`
- estimated cloud cost: `$125.83`
That is incompatible with any honest claim that Phase 4 has already been reached.
The same baseline also says:
- local mind: alive
- local session partner: usable
- local Hermes agent: not ready
So the repo's own truthful baseline says local capability exists, but zero-cloud operational sovereignty does not.
### 3. The model tracker is built to measure local-vs-cloud reality because the transition is not finished
Primary source:
- `metrics/model_tracker.py`
This file tracks:
- `local_sessions`
- `cloud_sessions`
- `local_pct`
- `est_cloud_cost`
- `est_saved`
That means the repo is architected to monitor a sovereignty transition, not to assume it is already complete.
### 4. There is already a proof harness — and its existence implies proof is still needed
Primary source:
- `scripts/local_timmy_proof_test.py`
This script explicitly searches for cloud/remote markers including:
- `chatgpt.com/backend-api/codex`
- `generativelanguage.googleapis.com`
- `api.groq.com`
- `143.198.27.163`
It also frames the output question as:
- is the active harness already local-only?
- why or why not?
A repo does not add a proof script like this if the zero-cloud cutover is already a settled fact.
### 5. The local subtree is stronger than the harness, but it is still only the target architecture
Primary sources:
- `LOCAL_Timmy_REPORT.md`
- `timmy-local/README.md`
`LOCAL_Timmy_REPORT.md` documents real local-first building blocks:
- local caching
- local Evennia world shell
- local ingestion pipeline
- prompt warming
Those are important Phase-4-aligned components.
But the broader repo still includes evidence of non-sovereign dependencies or remote references, such as:
- `scripts/evennia/bootstrap_local_evennia.py` defaulting operator email to `alexpaynex@gmail.com`
- `timmy-local/evennia/commands/tools.py` hardcoding `http://143.198.27.163:3000/...`
- `uni-wizard/tools/network_tools.py` hardcoding `GITEA_URL = "http://143.198.27.163:3000"`
- `uni-wizard/v2/task_router_daemon.py` defaulting `--gitea-url` to that same remote endpoint
These are not necessarily cloud inference dependencies, but they are still external dependency anchors inconsistent with the spirit of “No cloud. No rate limits. No one can turn it off.”
## Contradictions and Drift
### Contradiction A — local architecture exists, but repo evidence says cutover is incomplete
- `LOCAL_Timmy_REPORT.md` celebrates local infrastructure delivery.
- `reports/production/2026-03-29-local-timmy-baseline.md` still records `400 cloud` sessions and `0.7%` local.
These are not actually contradictory if read honestly:
- the local stack was delivered
- the fleet had not yet switched over to it
### Contradiction B — the local README was overstating current reality
Before this PR, `timmy-local/README.md` said the stack:
- “Runs entirely on your hardware with no cloud dependencies for core functionality.”
That sentence was too strong given the rest of the repo evidence:
- cloud defaults were still documented in the cutover plan
- cloud session volume was still quantified in the baseline report
- remote service references still existed across multiple scripts
This PR fixes that wording so the README describes `timmy-local` as the destination shape, not proof that the whole harness is already sovereign.
### Contradiction C — Phase 4 wants zero cloud dependencies, but the repo still documents explicit cloud-era markers
The repo itself still names or scans for:
- `openai-codex`
- `chatgpt.com/backend-api/codex`
- `generativelanguage.googleapis.com`
- `api.groq.com`
- `GROQ_API_KEY`
That does not mean the system can never become sovereign. It does mean the repo currently documents an unfinished migration boundary.
## Verdict
**Phase 4 is not yet reached.**
Why:
1. the repo's own baseline report still shows `403 total | 3 local | 400 cloud`
2. the repo's cutover plan still lists active cloud defaults and fallback paths as unresolved work
3. proof/guard scripts exist specifically to detect unresolved cloud and remote dependency markers
4. multiple runtime/ops files still point at external services such as `143.198.27.163`, `alexpaynex@gmail.com`, and Groq/OpenAI/Gemini-era paths
The truthful repo-grounded statement is:
- **local-first infrastructure exists**
- **zero-cloud sovereignty is the target**
- **the migration was not yet complete at the time this repo evidence was written**
## Highest-Leverage Next Actions
1. **Eliminate cloud defaults and hidden fallbacks first**
- follow `specs/2026-03-29-local-only-harness-cutover-plan.md`
- remove `openai-codex`, Gemini fallback, and any active cloud default path
2. **Kill cron inheritance bugs**
- no enabled cron should run with null model/provider if cloud defaults still exist anywhere
3. **Quarantine remote-ops scripts and hardcoded remote endpoints**
- `143.198.27.163` still appears in active repo scripts and command surfaces
- move legacy remote ops into quarantine or replace with local truth surfaces
4. **Run and preserve proof artifacts, not just intentions**
- the repo already has `scripts/local_timmy_proof_test.py`
- use it as the phase-gate proof generator
5. **Use the sovereignty scoreboard as a real gate**
- Phase 4 should not be declared complete while reports still show materially nonzero cloud sessions as the operating norm
## Definition of Done
Issue #551 should only be considered truly complete when the repo can point to evidence that all of the following are true:
1. no active model default points to a remote inference API
2. no fallback path silently escapes to cloud inference
3. no enabled cron can inherit a remote model/provider
4. active runtime paths no longer depend on Groq/OpenAI/Gemini-era inference markers
5. operator-critical services do not depend on external platforms like Gmail
6. remote hardcoded ops endpoints such as `143.198.27.163` are removed from the active Timmy path or clearly quarantined
7. the local proof script passes end-to-end
8. the sovereignty scoreboard shows cloud usage reduced to the point that “Zero Cloud Dependencies” is a truthful operational statement, not just an architectural aspiration
## Recommendation for This PR
This PR should **advance** Phase 4 by making the repo's public local-first docs honest and by recording a clear audit of why the milestone remains open.
That means the right PR reference style is:
- `Refs #551`
not:
- `Closes #551`
because the evidence in this repo shows the milestone is still in progress.
*Sovereignty and service always.*

View File

@@ -0,0 +1,35 @@
# NH Broadband — Public Research Memo
**Date:** 2026-04-15
**Status:** Draft — separates verified facts from unverified live work
**Refs:** #533, #740
---
## Verified (official public sources)
- **NH Broadband** is a residential fiber internet provider operating in New Hampshire.
- Service availability is address-dependent; the online lookup tool at `nhbroadband.com` reports coverage by street address.
- Residential fiber plans are offered; speed tiers vary by location.
- Scheduling line: **1-800-NHBB-INFO** (published on official site).
- Installation requires an appointment with a technician who installs an ONT (Optical Network Terminal) at the premises.
- Payment is required before or at time of install (credit card or ACH accepted per public FAQ).
## Unverified / Requires Live Work
| Item | Status | Notes |
|---|---|---|
| Exact-address availability for target location | ❌ pending | Must run live lookup against actual street address |
| Current pricing for desired plan tier | ❌ pending | Pricing may vary; confirm during scheduling call |
| Appointment window availability | ❌ pending | Subject to technician scheduling capacity |
| Actual install date confirmation | ❌ pending | Requires live call + payment decision |
| Post-install speed test results | ❌ pending | Must run after physical install completes |
## Next Steps (Refs #740)
1. Run address availability lookup on `nhbroadband.com`
2. Call 1-800-NHBB-INFO to schedule install
3. Confirm payment method
4. Receive appointment confirmation number
5. Prepare site (clear ONT install path)
6. Post-install: speed test and log results

View File

@@ -0,0 +1,132 @@
# Session Harvest Report — 2026-04-14
Date harvested: 2026-04-14
Prepared in repo: `Timmy_Foundation/timmy-home`
Verified against live forge state: 2026-04-15
Source issue: `timmy-home#648`
## Summary
This report turns the raw issue note in `#648` into a durable repository artifact.
The issue body captured a strong day of output across `hermes-agent` and `timmy-home`, but its status table had already drifted by verification time. The original note listed all delivered PRs as `Open`. Live Gitea state no longer matches that snapshot.
Most of the listed PRs are now closed, and three of the `timmy-home` PRs were merged successfully:
- PR #628
- PR #641
- PR #638
The rest of the delivered PRs are now `Closed (not merged)`.
This report preserves the harvest ledger while telling the truth about current forge state.
## Issue body drift
The issue body in `#648` is not wrong as a historical snapshot, but it is stale as an operational dashboard.
Verified changes since the original session note:
- every listed delivered PR is no longer open
- several blocked / skip items also changed state after the note was written
- the original `11 PRs open` framing no longer reflects current world state
That matters because this report is meant to be a harvest artifact, not a stale control panel.
## Delivered PR Ledger
### hermes-agent deliveries
| Work item | PR | Current forge state | Notes |
|-----------|----|---------------------|-------|
| hermes-agent #334 — Profile-scoped cron | PR #393 | Closed (not merged) | `feat(cron): Profile-scoped cron with parallel execution (#334)` |
| hermes-agent #251 — Memory contradiction detection | PR #413 | Closed (not merged) | `feat(memory): Periodic contradiction detection and resolution (#251)` |
| hermes-agent #468 — Cron cloud localhost warning | PR #500 | Closed (not merged) | `fix(cron): inject cloud-context warning when prompt refs localhost (#468)` |
| hermes-agent #499 — Hardcoded paths fix | PR #520 | Closed (not merged) | `fix: remove hardcoded ~/.hermes paths from optional skills (#499)` |
| hermes-agent #505 — Session templates | PR #553 | Closed (not merged) | `feat(templates): Session templates for code-first seeding (#505)` |
### timmy-home deliveries
| Work item | PR | Current forge state | Notes |
|-----------|----|---------------------|-------|
| timmy-home #590 — Emacs control plane | PR #624 | Closed (not merged) | `feat: Emacs Sovereign Control Plane (#590)` |
| timmy-home #587 — KTF processing log | PR #628 | Merged | `feat: Know Thy Father processing log and tracker (#587)` |
| timmy-home #583 — Phase 1 media indexing | PR #632 | Closed (not merged) | `feat: Know Thy Father Phase 1 — Media Indexing (#583)` |
| timmy-home #584 — Phase 2 analysis pipeline | PR #641 | Merged | `feat: Know Thy Father Phase 2 — Multimodal Analysis Pipeline (#584)` |
| timmy-home #579 — Ezra/Bezalel @mention fix | PR #635 | Closed (not merged) | `fix: VPS-native Gitea @mention heartbeat for Ezra/Bezalel (#579)` |
| timmy-home #578 — Big Brain Testament | PR #638 | Merged | `feat: Big Brain Testament rewrite artifact (#578)` |
## Triage Actions
The issue body recorded two triage actions:
- Closed #375 as stale (`deploy-crons.py` no longer exists)
- Triaged #510 with findings
Current forge state now verifies:
- #375 is closed
- #510 is also closed
So the reportable truth is that both triage actions are no longer pending. They are historical actions that have since resolved into closed issue state.
## Blocked / Skip Items
The issue body recorded three blocked / skip items:
- #511 Marathon guard — feature doesn't exist yet
- #556 `_classify_runtime` edge case — function doesn't exist in current codebase
- timmy-config #553557 a11y issues — reference Gitea frontend templates, not our repos
Verified current state for the `timmy-home` items:
- #511 remains open
- #556 is now closed
This means the blocked / skip section also drifted after harvest time.
Operationally accurate summary now:
- #511 remains open and unresolved from that blocked set
- #556 is no longer an active blocked item because it is closed
- the timmy-config accessibility note remains an external-scope observation rather than a `timmy-home` implementation item
## Current Totals
Verified from the 11 delivered PRs listed in the issue body:
- total PR artifacts harvested: 11
- current merged count: 3
- current closed-not-merged count: 8
- currently open count from this ledger: 0
So the current ledger is not:
- `11 open PRs`
It is:
- `3 merged`
- `8 closed without merge`
## Interpretation
This harvest still matters.
The value of the session is not only whether every listed PR merged. The value is that the work was surfaced, tracked, and moved into visible forge artifacts across multiple repos.
But the harvest report has to separate two things clearly:
1. what was produced on 2026-04-14
2. what is true on the forge now
That is why this artifact exists.
## Verification Method
The current report was verified by direct Gitea API reads against:
- `timmy-home#648`
- all PR numbers named in the issue body
- triage / blocked issue numbers #375, #510, #511, and #556
No unverified status claims are carried forward from the issue note without a live check.
## Bottom Line
The 2026-04-14 session produced a real harvest across `hermes-agent` and `timmy-home`.
But as of verification time, the exact truth is:
- the body of `#648` is a historical snapshot
- the snapshot drifted
- this report preserves the harvest while correcting the state ledger
That makes it useful as an ops artifact instead of just an old issue comment.

View File

@@ -0,0 +1,98 @@
# Burn Lane Empty Audit — timmy-home #662
Generated: 2026-04-17T03:42:50Z
Source issue: `[ops] Burn lane empty — all open issues triaged (2026-04-14)`
## Source Snapshot
Issue #662 is an operational status note, not a normal feature request. Its body is a historical snapshot of one burn lane claiming the queue was exhausted and recommending bulk closure of stale-open items.
## Live Summary
- Referenced issues audited: 42
- Already closed: 30
- Open but likely closure candidates (merged PR found): 1
- Open with active PRs: 0
- Open / needs manual review: 11
## Issue Body Drift
The body of #662 is not current truth. It mixes closed issues, open issues, ranges, and process notes into one static snapshot. This audit re-queries every referenced issue and classifies it against live forge state instead of trusting the original note.
| Issue | State | Classification | PR Summary |
|---|---|---|---|
| #579 | closed | already closed | closed PR #644, closed PR #643, closed PR #640, closed PR #635, closed PR #620 |
| #648 | open | needs manual review | closed PR #731 |
| #647 | closed | already closed | issue already closed |
| #619 | closed | already closed | issue already closed |
| #616 | closed | already closed | issue already closed |
| #614 | closed | already closed | issue already closed |
| #613 | closed | already closed | issue already closed |
| #660 | closed | already closed | issue already closed |
| #659 | closed | already closed | closed PR #660 |
| #658 | closed | already closed | issue already closed |
| #657 | closed | already closed | issue already closed |
| #656 | closed | already closed | closed PR #658 |
| #655 | closed | already closed | issue already closed |
| #654 | closed | already closed | closed PR #661 |
| #653 | closed | already closed | closed PR #660, closed PR #655 |
| #652 | closed | already closed | closed PR #660, merged PR #657, closed PR #655 |
| #651 | closed | already closed | closed PR #655 |
| #650 | closed | already closed | closed PR #661, closed PR #660, merged PR #654, closed PR #651 |
| #649 | closed | already closed | closed PR #660, merged PR #657, closed PR #651 |
| #646 | closed | already closed | closed PR #655, closed PR #651 |
| #582 | open | closure candidate | merged PR #641, merged PR #639, merged PR #637, merged PR #631, merged PR #630 |
| #627 | closed | already closed | issue already closed |
| #631 | closed | already closed | issue already closed |
| #632 | closed | already closed | issue already closed |
| #634 | closed | already closed | issue already closed |
| #639 | closed | already closed | issue already closed |
| #641 | closed | already closed | issue already closed |
| #575 | closed | already closed | closed PR #658, merged PR #656 |
| #576 | closed | already closed | merged PR #664, closed PR #663, closed PR #660, closed PR #655, merged PR #654, closed PR #651, closed PR #646, closed PR #642, closed PR #633 |
| #578 | closed | already closed | merged PR #638, closed PR #636 |
| #636 | closed | already closed | issue already closed |
| #638 | closed | already closed | issue already closed |
| #547 | open | needs manual review | closed PR #730 |
| #548 | open | needs manual review | closed PR #712 |
| #549 | open | needs manual review | closed PR #729 |
| #550 | open | needs manual review | closed PR #727 |
| #551 | open | needs manual review | closed PR #725 |
| #552 | open | needs manual review | closed PR #724 |
| #553 | open | needs manual review | closed PR #722 |
| #562 | open | needs manual review | closed PR #718 |
| #544 | open | needs manual review | closed PR #732 |
| #545 | open | needs manual review | closed PR #719 |
## Closure Candidates
These issues are still open but already have merged PR evidence in the forge and should be reviewed for bulk closure.
| Issue | State | Classification | PR Summary |
|---|---|---|---|
| #582 | open | closure candidate | merged PR #641, merged PR #639, merged PR #637, merged PR #631, merged PR #630 |
## Still Open / Needs Manual Review
These issues either have no matching PR signal or still have an active PR / ambiguous state and should stay in a human review lane.
| Issue | State | Classification | PR Summary |
|---|---|---|---|
| #648 | open | needs manual review | closed PR #731 |
| #547 | open | needs manual review | closed PR #730 |
| #548 | open | needs manual review | closed PR #712 |
| #549 | open | needs manual review | closed PR #729 |
| #550 | open | needs manual review | closed PR #727 |
| #551 | open | needs manual review | closed PR #725 |
| #552 | open | needs manual review | closed PR #724 |
| #553 | open | needs manual review | closed PR #722 |
| #562 | open | needs manual review | closed PR #718 |
| #544 | open | needs manual review | closed PR #732 |
| #545 | open | needs manual review | closed PR #719 |
## Recommendation
1. Close the `closure_candidate` issues in one deliberate ops pass after a final spot-check on main.
2. Leave `active_pr` items open until the current PRs are merged or closed.
3. Investigate `needs_manual_review` items individually — they may be report-only, assigned elsewhere, or still actionable.
4. Use this audit artifact instead of the raw body text of #662 for future lane-empty claims.

View File

@@ -0,0 +1,94 @@
# LAB-006 Call Log and Quote Template
Issue: #531
Purpose: capture the live calls, written confirmations, and septic cost options needed to close the issue honestly.
## County / State Call Log
### 1. NHDES Subsurface Systems Bureau
- Date:
- Time:
- Person reached:
- Phone used: 603-271-3501
- Email if follow-up requested: LRM-ARC@des.nh.gov
- Summary:
- Exact answer on whether a permitted designer is required for the 1-bedroom revision:
- Exact answer on whether owner-install is permitted for this parcel / use case:
- Exact answer on revision fee:
- Exact answer on whether moving the driveway triggers resubmission:
- Written follow-up promised? yes / no
- Reference number / email thread:
### 2. Local building / occupancy authority
- Date:
- Time:
- Office reached:
- Person reached:
- Phone:
- Summary:
- Does local occupancy sign-off require anything beyond NHDES septic approval?
- Separate permit / fee / inspection required?
- Written follow-up promised? yes / no
- Reference number / email thread:
### 3. Other agency / health / planning contact
- Date:
- Time:
- Office reached:
- Person reached:
- Phone:
- Summary:
- Key answer:
- Written follow-up promised? yes / no
- Reference number / email thread:
## Original Plan / Permit Retrieval Log
- Property address:
- Owner name searched:
- Approval number searched:
- OneStop searched? yes / no
- OneStop result:
- Archive request submitted? yes / no
- Archive request ID:
- Files received:
- Notes:
## Engineer / Designer Quote Tracker
| Vendor | Contact | Scope | Price | Lead time | Notes |
|---|---|---|---:|---|---|
| Designer 1 | | Revise approved plan to 1-bedroom | | | |
| Designer 2 | | Revise approved plan to 1-bedroom | | | |
| Designer 3 | | Revise approved plan to 1-bedroom | | | |
## Quote Tracker
| Option | Vendor / Person | Scope | Price | Lead time | Notes |
|---|---|---|---:|---|---|
| Professional install | | Full install | | | |
| Friend-with-excavator | | Excavation / install help | | | |
| Materials-only | | Tank + pipe + stone + misc. | | | |
## Materials List Draft
Use only if owner-install remains legally viable after the live calls.
- Septic tank:
- Distribution box:
- Pipe:
- Stone / leach field media:
- Fabric / protection:
- Inspection / riser components:
- Equipment rental:
- Delivery:
- Other:
## Final Yes / No Gate
- Revised 1-bedroom plan must be prepared by permitted designer: yes / no
- Owner-install permitted for this exact project: yes / no
- Revised plan fee confirmed: yes / no
- Local occupancy / building sign-off path confirmed: yes / no
- Three real quotes received: yes / no
- Best next action:

View File

@@ -0,0 +1,156 @@
# LAB-006 Septic Research
Issue: #531
Date: 2026-04-15
Status: public-doc research packet complete; live county/town calls and real quotes still pending
## Scope of this packet
This is a proof-oriented research packet built from public New Hampshire sources.
I did not claim any phone call, written county confirmation, engineer quote, or filed revision that did not actually happen.
What this packet does provide:
- official public source links
- a clearer answer on designer-vs-owner responsibilities
- the records lookup path for the existing approved septic plan
- the state contact point to call next
- a structured call and quote template for the live follow-up work
## Most important findings
### 1. A revised septic application in New Hampshire still appears to require a permitted designer
Official NHDES septic systems page:
- https://www.des.nh.gov/land/septic-systems
Direct language from the page:
- "Plans for proposed septic systems must be designed, prepared and submitted by a permitted New Hampshire septic system designer."
Implication for LAB-006:
- downsizing the approved plan from 3-4 bedroom to 1-bedroom is probably not a self-drawn paper edit if it changes the approved septic design/load assumptions
- moving the driveway on paper may also need designer involvement if it affects the approved layout or any required setback/field configuration
### 2. Owner-install appears possible in New Hampshire, but only in a narrow case
Official NHDES designer/installer page:
- https://www.des.nh.gov/land/septic-systems/septic-designer-or-installer
Direct language from the page:
- "Applications for individual sewage disposal systems or septic systems must be prepared by a permitted designer."
- "With the exception for homeowners installing for their primary domicile, septic systems must be constructed by a permitted installer."
Implication for LAB-006:
- public state guidance points to this answer:
- owner-install: likely YES, but only if the dwelling is the homeowner's primary domicile
- owner-designed / owner-submitted revised plan: public docs point to NO, because the application must still be prepared by a permitted designer
This is the strongest public answer I found without making the required phone calls.
### 3. The original approved septic documents should be searched in the NHDES records portal first
Official records portal / septic page:
- Septic records overview: https://www.des.nh.gov/land/septic-systems
- Subsurface OneStop portal: https://www4.des.state.nh.us/SSBOneStop/
Direct language from the septic systems page:
- "Our online Subsurface Onestop portal provides access to septic system records from 19671986 and 2016present. You can search by property owner name, address, designer, installer or approval number."
- "Records from 19862016 are currently being digitized."
- "If you cannot locate your septic record in the SSB Onestop Portal, you may submit an archive request online."
Implication for LAB-006:
- first check OneStop for the approved plan and approval number
- if the property falls into the digitization gap, file the archive request instead of guessing
### 4. Public docs point first to NHDES Subsurface Systems Bureau, not just a county office
Official contacts:
- NHDES Septic (Subsurface) forms portal: https://onlineforms.nh.gov/home/?Organizationcode=NHDES_Septic
- NHDES Contact page: https://www.des.nh.gov/contact
Public contact details shown in NHDES materials:
- Subsurface Systems Bureau phone: 603-271-3501
- LRM Application Receipt Center email: LRM-ARC@des.nh.gov
- Mailing address: NHDES Subsurface Systems Bureau, 29 Hazen Drive, PO Box 95, Concord, NH 03302-0095
Important note:
- the issue body says to call Sullivan County Building/Health
- the public New Hampshire septic program pages point to the state Subsurface Systems Bureau for the septic application/design side
- that does NOT prove the town/county has no role in occupancy or local building sign-off
- it does mean the next call should include NHDES, not only a county office
### 5. Revised forms are required as of February 1, 2026
Official septic systems page:
- https://www.des.nh.gov/land/septic-systems
Direct language:
- "Effective February 1, 2026: All submissions must comply with the revised Administrative Rules and use the revised forms."
Implication for LAB-006:
- if a revised plan is submitted, use the current NHDES septic forms rather than any old approval packet templates
## Public-source answer to the main yes/no question
Based on the public NHDES pages reviewed today:
- Can the owner revise and submit the septic plan without a designer?
- Public-doc answer: probably NO. The application/plans must be prepared by a permitted New Hampshire septic system designer.
- Can the owner install the septic system personally?
- Public-doc answer: possibly YES, but only for a homeowner installing for their primary domicile.
This is still not the same as county/town confirmation for this exact parcel and occupancy path. That call is still required.
## Best next live actions
1. Search the existing approval in Subsurface OneStop:
- by owner name
- by property address
- by designer name if known
- by approval number if any prior paperwork exists
2. If the file is not in OneStop, submit archive request.
3. Call NHDES Subsurface Systems Bureau at 603-271-3501 and ask:
- does downsizing an already-approved 3-4 bedroom septic plan to 1-bedroom require a newly prepared plan by a permitted designer?
- if the owner intends to self-install for a primary domicile, what exact homeowner-install form/process applies?
- what fee applies to revising an existing approved plan?
- does moving the driveway on the approved drawing trigger designer resubmission, site review, or other plan revision requirements?
4. Call the local building / occupancy authority for the parcel and confirm:
- who actually signs off the occupancy permit
- whether they defer fully to NHDES for septic revision
- whether any separate local building/driveway/site paperwork is required
5. If NHDES confirms designer-prepared revision is mandatory, get a designer quote immediately instead of spending more time on owner-submittal paths.
## What I did NOT verify
I did not verify any of the following as completed facts:
- that Sullivan County itself is the final septic approval authority for this parcel
- that a revised 1-bedroom plan has already been drafted or submitted
- that owner-install is permitted for this exact property after all local conditions are applied
- the exact revision fee
- any real contractor quote
## Recommended practical interpretation
Todays public-doc evidence strongly supports this working assumption:
- design/revision work -> permitted septic designer
- physical installation -> homeowner may be able to do it for a primary domicile
- records/process/questions -> start with NHDES Subsurface Systems Bureau and OneStop
That is enough to stop guessing and start the right calls.
## Evidence links
- NHDES Septic Systems: https://www.des.nh.gov/land/septic-systems
- NHDES Septic Designer and Installer: https://www.des.nh.gov/land/septic-systems/septic-designer-or-installer
- NHDES Septic Online Forms: https://onlineforms.nh.gov/home/?Organizationcode=NHDES_Septic
- NHDES Subsurface OneStop: https://www4.des.state.nh.us/SSBOneStop/
- NHDES Contact page: https://www.des.nh.gov/contact
## Deliverables in this PR
- this research memo
- a call-log and quote-tracker template for the live follow-up work

View File

@@ -0,0 +1,218 @@
# QA Triage Action Plan — Foundation-Wide (2026-04-14)
> **Source:** Issue #691 — Cross-Repo Deep QA Report
> **Generated:** 2026-04-14
> **Status:** Active triage — actionable steps for each finding
---
## Executive Summary
The QA sweep identified systemic issues across the Foundation. Current state (verified live):
| Metric | QA Report | Current | Trend |
|--------|-----------|---------|-------|
| Total open PRs | ~55+ | **166** | Worsening |
| Repos with dupes | 3 | **5 (all)** | Worsening |
| Duplicate PR issues | 7+ | **58** | Critical |
| Prod surfaces reachable | 0/4 | 0/4 | Unchanged |
**The core problem:** Burn sessions generate faster than triage can absorb. The backlog is growing, not shrinking.
---
## P0 — Critical
### 1. Production Surfaces Down (404 on all endpoints)
**Status:** Unchanged since QA report
**Impact:** Zero users can reach any Timmy surface. The Door (crisis intervention) is unreachable.
| Surface | URL | Status |
|---------|-----|--------|
| Root | http://143.198.27.163/ | nginx 404 |
| Nexus | http://143.198.27.163/nexus/ | 404 |
| Playground | http://143.198.27.163/playground/ | 404 |
| Tower | http://143.198.27.163/tower/ | 404 |
| Domain | https://alexanderwhitestone.com/ | DNS broken |
**Action:**
- [ ] Verify DNS records for alexanderwhitestone.com (check registrar)
- [ ] SSH to VPS, check nginx config: `nginx -T`
- [ ] Ensure server blocks exist for each location
- [ ] Restart nginx: `systemctl restart nginx`
- [ ] Tracked in the-nexus#1105
**Owner:** Infrastructure
**Priority:** Immediate — this is the mission
### 2. the-playground index.html Broken
**Status:** Unconfirmed since QA report
**Impact:** Playground app crashes on load — missing script tags
**Action:**
- [ ] Read the-playground/index.html
- [ ] Verify script tags for all JS modules
- [ ] Fix missing imports
- [ ] Tracked in the-playground#200
**Owner:** the-playground
**Priority:** High — blocks user-facing playground
---
## P1 — High (Duplicate PR Crisis)
### 3. Duplicate PR Storm Across All Repos
**Current state (verified live 2026-04-14):**
| Repo | Open PRs | Issues with Duplicates | Worst Case |
|------|----------|----------------------|------------|
| the-nexus | 44 | 16 | Issue #1509 → 4 PRs |
| the-playground | 31 | 10 | Issue #180 → 3 PRs |
| the-door | 27 | 6 | Issue #988 → 7 PRs |
| timmy-config | 50 | 20 | Issue #50 → 7 PRs |
| timmy-home | 14 | 6 | Issue #50 → 6 PRs |
| **Total** | **166** | **58 issues** | — |
**Root cause:** Burn sessions create branches without checking for existing PRs on the same issue. No deduplication gate in the burn pipeline.
**Immediate action — close duplicates per repo:**
For each issue with multiple PRs:
1. Keep the PR with the most commits/diff (most complete implementation)
2. Close all others with comment: "Closing duplicate. See #PR for primary implementation."
3. If no PR is clearly superior, keep the oldest (first mover)
**Script to identify duplicates:**
```bash
# For each repo, list issues with >1 open PR
python3 scripts/duplicate-pr-detector.py --repo <repo> --close-duplicates
```
**Long-term fix:**
- [ ] Add pre-flight check to burn loop: query open PRs before creating new branch
- [ ] Add Gitea label `burn-active` to track which issues have active burn PRs
- [ ] Add CI check that rejects PR if another open PR references the same issue
**Owner:** Fleet / Burn infrastructure
**Priority:** High — duplicates waste review time and create merge conflicts
### 4. Misfiled PR in wrong repo
**the-nexus PR #1521:** "timmy-home Backlog Triage Report" is filed in the-nexus but concerns timmy-home.
**Action:**
- [ ] Close PR #1521 in the-nexus with redirect comment
- [ ] File content as issue or PR in timmy-home if still relevant
---
## P2 — Medium
### 5. the-door Crisis Features Blocked
Mission-critical PRs sitting unreviewed:
| Issue | Title | Impact |
|-------|-------|--------|
| #91 | Safety plan improvements | User safety |
| #89 | Safety plan enhancements | User safety |
| #90 | Crisis overlay fixes | UX |
| #87 | Crisis overlay bugs | UX |
| 988 link | Crisis hotline link fix | **Life safety** |
**Action:**
- [ ] Prioritize the-door PR review over all other repos
- [ ] Assign a reviewer or run dedicated triage session for the-door only
- [ ] After review, merge in dependency order
**Owner:** Crisis team / Alexander
**Priority:** High — this is the mission
### 6. Branch Protection Missing Foundation-Wide
No repo has branch protection enabled. Any member can push directly to main.
**Action:**
- [ ] Enable branch protection on all repos with:
- Require 1 approval before merge
- Require CI to pass (where CI exists)
- Dismiss stale approvals on new commits
- [ ] Covered in timmy-home PR #606 but not yet implemented
**Repos without CI (need smoke test first):**
- the-playground
- the-beacon
- timmy-home
**Owner:** Alexander / Infrastructure
**Priority:** Medium — prevents accidental breakage
---
## P3 — Low (Process Improvements)
### 7. Burn Session Deduplication Gate
**Problem:** Burn loops don't check for existing PRs before creating new ones.
**Solution:** Pre-flight check in burn pipeline:
```python
def has_open_pr(owner, repo, issue_number):
prs = gitea.get_pulls(owner, repo, state="open")
for pr in prs:
if f"#{issue_number}" in (pr.get("body", "") or ""):
return True
return False
```
**Action:**
- [ ] Add to hermes-agent burn loop
- [ ] Add to timmy-config burn scripts
- [ ] Test with dry-run before enabling
### 8. Nightly Triage Cron
**Problem:** No automated triage. Duplicates accumulate until manual sweep.
**Solution:** Nightly cron that:
1. Scans all repos for duplicate PRs
2. Posts summary to a triage channel
3. Auto-closes duplicates older than 48h with lower diff count
**Action:**
- [ ] Design triage cron job spec
- [ ] Implement as hermes cron job
- [ ] Run nightly at 03:00 UTC
---
## Priority Order (Execution Sequence)
1. **Fix DNS/nginx** — The Door must be reachable (crisis intervention = the mission)
2. **Close duplicate PRs** — 58 issues with dupes, clear the noise
3. **Review the-door PRs** — Mission-critical crisis features
4. **Fix the-playground** — User-facing app broken
5. **Enable branch protection** — Prevent future breakage
6. **Build dedup gate** — Prevent future duplicate storms
7. **Nightly triage cron** — Automated hygiene
---
## Verification Checklist
After completing actions above, verify:
- [ ] http://143.198.27.163/ returns a page (not 404)
- [ ] https://alexanderwhitestone.com/ resolves
- [ ] All repos have <5 duplicate PRs
- [ ] the-door has 0 unreviewed safety/crisis PRs
- [ ] Branch protection enabled on all repos
- [ ] Burn loop has pre-flight PR check
---
*This plan converts QA findings into executable actions. Each item has an owner, priority, and verification step.*

View File

@@ -0,0 +1,56 @@
# Triage Cadence Report — timmy-home (2026-04-15)
> Issue #685 | Backlog reduced from 220 to 50
## Summary
timmy-home's open issue count dropped from 220 (peak) to 50 through batch-pipeline codebase genome generation and triage. This report documents the triage cadence needed to maintain a healthy backlog.
## Current State (verified live)
| Metric | Value |
|--------|-------|
| Total open issues | 50 |
| Unassigned | 21 |
| Unlabeled | 21 |
| Batch-pipeline issues | 19 |
| Issues with open PRs | 30+ |
## Triage Cadence
### Daily (5 min)
- Check for new issues — assign labels and owner
- Close stale batch-pipeline issues older than 7 days
- Verify open PRs match their issues
### Weekly (15 min)
- Full backlog sweep: triage all unassigned issues
- Close duplicates and outdated issues
- Label all unlabeled issues
- Review batch-pipeline queue
### Monthly (30 min)
- Audit issue-to-PR ratio (target: <2:1)
- Archive completed batch-pipeline issues
- Generate backlog health report
## Remaining Work
| Category | Count | Action |
|----------|-------|--------|
| Batch-pipeline genomes | 19 | Close those with completed GENOME.md PRs |
| Unassigned | 21 | Assign or close |
| Unlabeled | 21 | Add labels |
| No PR | ~20 | Triage or close |
## Recommended Labels
- `batch-pipeline` — Auto-generated pipeline issues
- `genome` — Codebase genome analysis
- `ops` — Operations/infrastructure
- `documentation` — Docs and reports
- `triage` — Needs triage
---
*Generated: 2026-04-15 | timmy-home issue #685*

View File

@@ -0,0 +1,102 @@
# Long Context vs RAG Decision Framework
**Research Backlog Item #4.3** | Impact: 4 | Effort: 1 | Ratio: 4.0
**Date**: 2026-04-15
**Status**: RESEARCHED
## Executive Summary
Modern LLMs have 128K-200K+ context windows, but we still treat them like 4K models by default. This document provides a decision framework for when to stuff context vs. use RAG, based on empirical findings and our stack constraints.
## The Core Insight
**Long context ≠ better answers.** Research shows:
- "Lost in the Middle" effect: Models attend poorly to information in the middle of long contexts (Liu et al., 2023)
- RAG with reranking outperforms full-context stuffing for document QA when docs > 50K tokens
- Cost scales quadratically with context length (attention computation)
- Latency increases linearly with input length
**RAG ≠ always better.** Retrieval introduces:
- Recall errors (miss relevant chunks)
- Precision errors (retrieve irrelevant chunks)
- Chunking artifacts (splitting mid-sentence)
- Additional latency for embedding + search
## Decision Matrix
| Scenario | Context Size | Recommendation | Why |
|----------|-------------|---------------|-----|
| Single conversation (< 32K) | Small | **Stuff everything** | No retrieval overhead, full context available |
| 5-20 documents, focused query | 32K-128K | **Hybrid** | Key docs in context, rest via RAG |
| Large corpus search | > 128K | **Pure RAG + reranking** | Full context impossible, must retrieve |
| Code review (< 5 files) | < 32K | **Stuff everything** | Code needs full context for understanding |
| Code review (repo-wide) | > 128K | **RAG with file-level chunks** | Files are natural chunk boundaries |
| Multi-turn conversation | Growing | **Hybrid + compression** | Keep recent turns in full, compress older |
| Fact retrieval | Any | **RAG** | Always faster to search than read everything |
| Complex reasoning across docs | 32K-128K | **Stuff + chain-of-thought** | Models need all context for cross-doc reasoning |
## Our Stack Constraints
### What We Have
- **Cloud models**: 128K-200K context (OpenRouter providers)
- **Local Ollama**: 8K-32K context (Gemma-4 default 8192)
- **Hermes fact_store**: SQLite FTS5 full-text search
- **Memory**: MemPalace holographic embeddings
- **Session context**: Growing conversation history
### What This Means
1. **Cloud sessions**: We CAN stuff up to 128K but SHOULD we? Cost and latency matter.
2. **Local sessions**: MUST use RAG for anything beyond 8K. Long context not available.
3. **Mixed fleet**: Need a routing layer that decides per-session.
## Advanced Patterns
### 1. Progressive Context Loading
Don't load everything at once. Start with RAG, then stuff additional docs as needed:
```
Turn 1: RAG search → top 3 chunks
Turn 2: Model asks "I need more context about X" → stuff X
Turn 3: Model has enough → continue
```
### 2. Context Budgeting
Allocate context budget across components:
```
System prompt: 2,000 tokens (always)
Recent messages: 10,000 tokens (last 5 turns)
RAG results: 8,000 tokens (top chunks)
Stuffed docs: 12,000 tokens (key docs)
---------------------------
Total: 32,000 tokens (fits 32K model)
```
### 3. Smart Compression
Before stuffing, compress older context:
- Summarize turns older than 10
- Remove tool call results (keep only final outputs)
- Deduplicate repeated information
- Use structured representations (JSON) instead of prose
## Empirical Benchmarks Needed
1. **Stuffing vs RAG accuracy** on our fact_store queries
2. **Latency comparison** at 32K, 64K, 128K context
3. **Cost per query** for cloud models at various context sizes
4. **Local model behavior** when pushed beyond rated context
## Recommendations
1. **Audit current context usage**: How many sessions hit > 32K? (Low effort, high value)
2. **Implement ContextRouter**: ~50 LOC, adds routing decisions to hermes
3. **Add context-size logging**: Track input tokens per session for data gathering
## References
- Liu et al. "Lost in the Middle: How Language Models Use Long Contexts" (2023) — https://arxiv.org/abs/2307.03172
- Shi et al. "Large Language Models are Easily Distracted by Irrelevant Context" (2023)
- Xu et al. "Retrieval Meets Long Context LLMs" (2023) — hybrid approaches outperform both alone
- Anthropic's Claude 3.5 context caching — built-in prefix caching reduces cost for repeated system prompts
---
*Sovereignty and service always.*

View File

@@ -1,46 +1,90 @@
# Big Brain Pod Verification
# Big Brain Provider Verification
Verification script for Big Brain pod with gemma3:27b model.
Repo wiring for the `big_brain` provider used by Mac Hermes.
## Issue #573
## Issue #543
[BIG-BRAIN] Verify pod live: gemma3:27b pulled and responding
[PROVE-IT] Timmy: Wire RunPod/Vertex AI Gemma 4 to Mac Hermes
## Pod Details
## What this repo now supports
- Pod ID: `8lfr3j47a5r3gn`
- GPU: L40S 48GB
- Image: `ollama/ollama:latest`
- Endpoint: `https://8lfr3j47a5r3gn-11434.proxy.runpod.net`
- Cost: $0.79/hour
The repo no longer hardcodes one dead RunPod pod as the truth.
Instead, it defines a **Big Brain provider contract**:
- provider name: `Big Brain`
- model: `gemma4:latest`
- endpoint style: OpenAI-compatible `/v1` by default
- verification path: `scripts/verify_big_brain.py`
## Verification Script
Supported deployment shapes:
1. **RunPod + Ollama/OpenAI-compatible bridge**
- Example base URL: `https://<pod-id>-11434.proxy.runpod.net/v1`
2. **Vertex AI through an OpenAI-compatible bridge/proxy**
- Example base URL: `https://<your-bridge-host>/v1`
`scripts/verify_big_brain.py` checks:
## Config wiring
1. `/api/tags` - Verifies gemma3:27b is in model list
2. `/api/generate` - Tests response time (< 30s requirement)
3. Uptime logging for cost awareness
`config.yaml` now carries a generic provider block:
```yaml
- name: Big Brain
base_url: https://YOUR_BIG_BRAIN_HOST/v1
api_key: ''
model: gemma4:latest
```
Override at runtime if needed:
- `BIG_BRAIN_BASE_URL`
- `BIG_BRAIN_MODEL`
- `BIG_BRAIN_BACKEND` (`openai` or `ollama`)
- `BIG_BRAIN_API_KEY`
## Verification scripts
### 1. `scripts/verify_big_brain.py`
Checks the configured provider using the right protocol for the chosen backend.
For `openai` backends it verifies:
- `GET /models`
- `POST /chat/completions`
For `ollama` backends it verifies:
- `GET /api/tags`
- `POST /api/generate`
Writes:
- `big_brain_verification.json`
### 2. `scripts/big_brain_manager.py`
A more verbose wrapper over the same provider contract.
Writes:
- `pod_verification_results.json`
## Usage
```bash
cd scripts
python3 verify_big_brain.py
python3 scripts/verify_big_brain.py
python3 scripts/big_brain_manager.py
```
## Output
## Honest current state
- Console output with verification results
- `big_brain_verification.json` with detailed results
- Exit code 0 on success, 1 on failure
On fresh main before this fix, the repo was pointing at a stale RunPod endpoint:
- `https://8lfr3j47a5r3gn-11434.proxy.runpod.net`
- verification returned HTTP 404 for both model listing and generation
## Acceptance Criteria
That meant the repo claimed Big Brain wiring existed, but the proof path was stale and tied to a dead specific pod.
- [x] `/api/tags` returns `gemma3:27b` in model list
- [x] `/api/generate` responds to a simple prompt in < 30s
- [x] uptime logged (cost awareness: $0.79/hr)
This fix makes the repo wiring reusable and truthful, but it does **not** provision a fresh paid GPU automatically.
## Previous Issues
## Acceptance mapping
Previous pod (elr5vkj96qdplf) used broken `runpod/ollama:latest` image and never started. Fix: use `ollama/ollama:latest`. Volume mount at `/root/.ollama` for model persistence.
What this repo change satisfies:
- [x] Mac Hermes has a `big_brain` provider contract in `config.yaml`
- [x] Verification script checks that provider through the same API shape Hermes needs
- [x] RunPod and Vertex-style wiring are documented without hardcoding a dead pod
What still depends on live infrastructure outside the repo:
- [ ] GPU instance actually provisioned and running
- [ ] endpoint responsive right now
- [ ] live `hermes chat --provider big_brain` success against a real endpoint

191
scripts/agent_pr_gate.py Executable file
View File

@@ -0,0 +1,191 @@
#!/usr/bin/env python3
import argparse
import json
import os
import re
import sys
import urllib.request
from pathlib import Path
API_BASE = "https://forge.alexanderwhitestone.com/api/v1"
LOW_RISK_PREFIXES = (
'docs/', 'reports/', 'notes/', 'tickets/', 'research/', 'briefings/',
'twitter-archive/notes/', 'tests/'
)
LOW_RISK_SUFFIXES = {'.md', '.txt', '.jsonl'}
MEDIUM_RISK_PREFIXES = ('.gitea/workflows/',)
HIGH_RISK_PREFIXES = (
'scripts/', 'deploy/', 'infrastructure/', 'metrics/', 'heartbeat/',
'wizards/', 'evennia/', 'uniwizard/', 'uni-wizard/', 'timmy-local/',
'evolution/'
)
HIGH_RISK_SUFFIXES = {'.py', '.sh', '.ini', '.service'}
def read_changed_files(path):
return [line.strip() for line in Path(path).read_text(encoding='utf-8').splitlines() if line.strip()]
def classify_risk(files):
if not files:
return 'high'
level = 'low'
for file_path in files:
path = file_path.strip()
suffix = Path(path).suffix.lower()
if path.startswith(LOW_RISK_PREFIXES):
continue
if path.startswith(HIGH_RISK_PREFIXES) or suffix in HIGH_RISK_SUFFIXES:
return 'high'
if path.startswith(MEDIUM_RISK_PREFIXES):
level = 'medium'
continue
if path.startswith(LOW_RISK_PREFIXES) or suffix in LOW_RISK_SUFFIXES:
continue
level = 'high'
return level
def validate_pr_body(title, body):
details = []
combined = f"{title}\n{body}".strip()
if not re.search(r'#\d+', combined):
details.append('PR body/title must include an issue reference like #562.')
if not re.search(r'(^|\n)\s*(verification|tests?)\s*:', body, re.IGNORECASE):
details.append('PR body must include a Verification: section.')
return (len(details) == 0, details)
def build_comment_body(syntax_status, tests_status, criteria_status, risk_level):
statuses = {
'syntax': syntax_status,
'tests': tests_status,
'criteria': criteria_status,
}
all_clean = all(value == 'success' for value in statuses.values())
action = 'auto-merge' if all_clean and risk_level == 'low' else 'human review'
lines = [
'## Agent PR Gate',
'',
'| Check | Status |',
'|-------|--------|',
f"| Syntax / parse | {syntax_status} |",
f"| Test suite | {tests_status} |",
f"| PR criteria | {criteria_status} |",
f"| Risk level | {risk_level} |",
'',
]
failed = [name for name, value in statuses.items() if value != 'success']
if failed:
lines.append('### Failure details')
for name in failed:
lines.append(f'- {name} reported failure. Inspect the workflow logs for that step.')
else:
lines.append('All automated checks passed.')
lines.extend([
'',
f'Recommendation: {action}.',
'Low-risk documentation/test-only PRs may be auto-merged. Operational changes stay in human review.',
])
return '\n'.join(lines)
def _read_event(event_path):
data = json.loads(Path(event_path).read_text(encoding='utf-8'))
pr = data.get('pull_request') or {}
repo = (data.get('repository') or {}).get('full_name') or os.environ.get('GITHUB_REPOSITORY')
pr_number = pr.get('number') or data.get('number')
title = pr.get('title') or ''
body = pr.get('body') or ''
return repo, pr_number, title, body
def _request_json(method, url, token, payload=None):
data = None if payload is None else json.dumps(payload).encode('utf-8')
headers = {'Authorization': f'token {token}', 'Content-Type': 'application/json'}
req = urllib.request.Request(url, data=data, headers=headers, method=method)
with urllib.request.urlopen(req, timeout=30) as resp:
return json.loads(resp.read().decode('utf-8'))
def post_comment(repo, pr_number, token, body):
url = f'{API_BASE}/repos/{repo}/issues/{pr_number}/comments'
return _request_json('POST', url, token, {'body': body})
def merge_pr(repo, pr_number, token):
url = f'{API_BASE}/repos/{repo}/pulls/{pr_number}/merge'
return _request_json('POST', url, token, {'Do': 'merge'})
def cmd_classify_risk(args):
files = list(args.files or [])
if args.files_file:
files.extend(read_changed_files(args.files_file))
print(json.dumps({'risk': classify_risk(files), 'files': files}, indent=2))
return 0
def cmd_validate_pr(args):
_, _, title, body = _read_event(args.event_path)
ok, details = validate_pr_body(title, body)
if ok:
print('PR body validation passed.')
return 0
for detail in details:
print(detail)
return 1
def cmd_comment(args):
repo, pr_number, _, _ = _read_event(args.event_path)
body = build_comment_body(args.syntax, args.tests, args.criteria, args.risk)
post_comment(repo, pr_number, args.token, body)
print(f'Commented on PR #{pr_number} in {repo}.')
return 0
def cmd_merge(args):
repo, pr_number, _, _ = _read_event(args.event_path)
merge_pr(repo, pr_number, args.token)
print(f'Merged PR #{pr_number} in {repo}.')
return 0
def build_parser():
parser = argparse.ArgumentParser(description='Agent PR CI helpers for timmy-home.')
sub = parser.add_subparsers(dest='command', required=True)
classify = sub.add_parser('classify-risk')
classify.add_argument('--files-file')
classify.add_argument('files', nargs='*')
classify.set_defaults(func=cmd_classify_risk)
validate = sub.add_parser('validate-pr')
validate.add_argument('--event-path', required=True)
validate.set_defaults(func=cmd_validate_pr)
comment = sub.add_parser('comment')
comment.add_argument('--event-path', required=True)
comment.add_argument('--token', required=True)
comment.add_argument('--syntax', required=True)
comment.add_argument('--tests', required=True)
comment.add_argument('--criteria', required=True)
comment.add_argument('--risk', required=True)
comment.set_defaults(func=cmd_comment)
merge = sub.add_parser('merge')
merge.add_argument('--event-path', required=True)
merge.add_argument('--token', required=True)
merge.set_defaults(func=cmd_merge)
return parser
def main(argv=None):
parser = build_parser()
args = parser.parse_args(argv)
return args.func(args)
if __name__ == '__main__':
sys.exit(main())

138
scripts/audit_trail.py Executable file
View File

@@ -0,0 +1,138 @@
#!/usr/bin/env python3
# audit_trail.py - Local logging of inputs, sources, and confidence.
# Implements SOUL.md "What Honesty Requires" - The Audit Trail.
# Logs are stored locally. Never sent anywhere. The user owns them.
# Part of #794
import json
import hashlib
import os
import time
from datetime import datetime, timezone
from pathlib import Path
from typing import Any, Dict, List, Optional
from dataclasses import dataclass, field, asdict
AUDIT_DIR = Path(os.environ.get("HERMES_HOME", Path.home() / ".hermes")) / "audit-trail"
@dataclass
class AuditEntry:
id: str
ts: str
input_text: str
sources: List[str]
confidence: float
output_text: str
model: str
provider: str = ""
session_id: str = ""
source_types: List[str] = field(default_factory=list)
@staticmethod
def generate_id(input_text: str, output_text: str, ts: str) -> str:
content = f"{ts}:{input_text}:{output_text}"
return hashlib.sha256(content.encode()).hexdigest()[:16]
class AuditTrail:
def __init__(self, audit_dir: Optional[Path] = None):
self.audit_dir = audit_dir or AUDIT_DIR
self.audit_dir.mkdir(parents=True, exist_ok=True)
self._log_file = self.audit_dir / "trail.jsonl"
def log_response(self, input_text, sources, confidence, output_text,
model="", provider="", session_id="", source_types=None):
ts = datetime.now(timezone.utc).isoformat()
entry = AuditEntry(
id=AuditEntry.generate_id(input_text, output_text, ts),
ts=ts,
input_text=input_text[:1000],
sources=[s[:200] for s in sources[:10]],
confidence=round(confidence, 3),
output_text=output_text[:2000],
model=model, provider=provider, session_id=session_id,
source_types=source_types or [],
)
with open(self._log_file, "a") as f:
f.write(json.dumps(asdict(entry)) + "\n")
return entry
def query(self, search_text, limit=10, min_confidence=0.0):
if not self._log_file.exists():
return []
results = []
search_lower = search_text.lower()
with open(self._log_file) as f:
for line in f:
line = line.strip()
if not line:
continue
try:
data = json.loads(line)
except json.JSONDecodeError:
continue
if data.get("confidence", 0) < min_confidence:
continue
searchable = (data.get("input_text", "") + " " +
data.get("output_text", "") + " " +
" ".join(data.get("sources", []))).lower()
if search_lower in searchable:
results.append(AuditEntry(**{k: data.get(k, "") if isinstance(data.get(k), str)
else data.get(k, []) if isinstance(data.get(k), list)
else data.get(k, 0.0) for k in AuditEntry.__dataclass_fields__}))
if len(results) >= limit:
break
return results
def get_stats(self):
if not self._log_file.exists():
return {"total": 0, "avg_confidence": 0, "sources_breakdown": {}}
total = 0
confidence_sum = 0.0
source_types = {}
with open(self._log_file) as f:
for line in f:
try:
data = json.loads(line.strip())
total += 1
confidence_sum += data.get("confidence", 0)
for st in data.get("source_types", []):
source_types[st] = source_types.get(st, 0) + 1
except (json.JSONDecodeError, ValueError):
continue
return {"total": total, "avg_confidence": round(confidence_sum / max(total, 1), 3),
"sources_breakdown": source_types}
def get_by_session(self, session_id, limit=50):
if not self._log_file.exists():
return []
results = []
with open(self._log_file) as f:
for line in f:
try:
data = json.loads(line.strip())
if data.get("session_id") == session_id:
results.append(AuditEntry(**{k: data.get(k, "") if isinstance(data.get(k), str)
else data.get(k, []) if isinstance(data.get(k), list)
else data.get(k, 0.0) for k in AuditEntry.__dataclass_fields__}))
except (json.JSONDecodeError, ValueError):
continue
if len(results) >= limit:
break
return results
_default_trail = None
def get_trail():
global _default_trail
if _default_trail is None:
_default_trail = AuditTrail()
return _default_trail
def log_response(**kwargs):
return get_trail().log_response(**kwargs)
def query(search_text, **kwargs):
return get_trail().query(search_text, **kwargs)

View File

@@ -0,0 +1,329 @@
#!/usr/bin/env python3
"""Create or refresh fleet incidents on Gitea from local infrastructure signals.
Refs: timmy-home #553
"""
from __future__ import annotations
import argparse
import json
import os
from dataclasses import dataclass
from datetime import datetime, timezone
from pathlib import Path
from typing import Iterable
from urllib import request
DEFAULT_BASE_URL = "https://forge.alexanderwhitestone.com/api/v1"
DEFAULT_OWNER = "Timmy_Foundation"
DEFAULT_REPO = "timmy-home"
DEFAULT_TOKEN_FILE = Path.home() / ".config" / "gitea" / "token"
DEFAULT_FAILOVER_STATUS = Path.home() / ".timmy" / "failover_status.json"
DEFAULT_RESTART_STATE_DIR = Path("/var/lib/timmy/restarts")
DEFAULT_HEARTBEAT_FILE = Path("/var/lib/timmy/heartbeats/fleet_health.last")
@dataclass(frozen=True)
class Incident:
fingerprint: str
title: str
body: str
def latest_evidence(self) -> str:
lines = [line for line in self.body.splitlines() if line.strip()]
if lines and lines[0].startswith("Fingerprint: "):
lines = lines[1:]
return "\n".join(lines).strip()
class GiteaClient:
def __init__(self, token: str, owner: str = DEFAULT_OWNER, repo: str = DEFAULT_REPO, base_url: str = DEFAULT_BASE_URL):
self.token = token
self.owner = owner
self.repo = repo
self.base_url = base_url.rstrip("/")
def _request(self, path: str, *, method: str = "GET", data: dict | None = None):
payload = None if data is None else json.dumps(data).encode()
headers = {"Authorization": f"token {self.token}"}
if payload is not None:
headers["Content-Type"] = "application/json"
req = request.Request(f"{self.base_url}{path}", data=payload, headers=headers, method=method)
with request.urlopen(req, timeout=30) as resp:
return json.loads(resp.read().decode())
def list_open_issues(self):
issues = self._request(f"/repos/{self.owner}/{self.repo}/issues?state=open&limit=100")
return [issue for issue in issues if not issue.get("pull_request")]
def create_issue(self, title: str, body: str):
return self._request(
f"/repos/{self.owner}/{self.repo}/issues",
method="POST",
data={"title": title, "body": body},
)
def comment_issue(self, issue_number: int, body: str):
return self._request(
f"/repos/{self.owner}/{self.repo}/issues/{issue_number}/comments",
method="POST",
data={"body": body},
)
def load_json(path: Path):
if not path.exists():
return None
return json.loads(path.read_text())
def load_restart_counts(state_dir: Path) -> dict[str, int]:
if not state_dir.exists():
return {}
counts: dict[str, int] = {}
for path in sorted(state_dir.glob("*.count")):
try:
counts[path.stem] = int(path.read_text().strip())
except ValueError:
continue
return counts
def heartbeat_is_stale(path: Path, *, now: datetime | None = None, max_age_seconds: int = 900) -> bool:
if now is None:
now = datetime.now(timezone.utc)
if not path.exists():
return True
age = now.timestamp() - path.stat().st_mtime
return age > max_age_seconds
def _iso(dt: datetime) -> str:
return dt.astimezone(timezone.utc).isoformat().replace("+00:00", "Z")
def _build_body(fingerprint: str, *details: str) -> str:
detail_lines = [detail for detail in details if detail]
return "\n".join([f"Fingerprint: {fingerprint}", *detail_lines])
def build_incidents(
*,
failover_status: dict | None,
restart_counts: dict[str, int],
heartbeat_stale: bool,
now: datetime | None = None,
restart_escalation_threshold: int = 3,
) -> list[Incident]:
if now is None:
now = datetime.now(timezone.utc)
incidents: list[Incident] = []
failover_timestamp = None
fleet = {}
if failover_status:
failover_timestamp = failover_status.get("timestamp")
fleet = failover_status.get("fleet") or {}
for host, status in sorted(fleet.items()):
if str(status).upper() == "ONLINE":
continue
fingerprint = f"host-offline:{host}"
failover_detail = f"Failover status timestamp: {failover_timestamp}" if failover_timestamp is not None else "Failover status timestamp: unknown"
incidents.append(
Incident(
fingerprint=fingerprint,
title=f"[AUTO] Fleet host offline: {host}",
body=_build_body(
fingerprint,
f"Detected at: {_iso(now)}",
failover_detail,
f"Host `{host}` reported `{status}` by failover monitor.",
),
)
)
for process_name, count in sorted(restart_counts.items()):
if count <= restart_escalation_threshold:
continue
fingerprint = f"restart-escalation:{process_name}"
incidents.append(
Incident(
fingerprint=fingerprint,
title=f"[AUTO] Restart escalation: {process_name}",
body=_build_body(
fingerprint,
f"Detected at: {_iso(now)}",
f"Process `{process_name}` has crossed the restart escalation threshold with count={count}.",
),
)
)
if heartbeat_stale:
fingerprint = "probe-stale:fleet-health"
incidents.append(
Incident(
fingerprint=fingerprint,
title="[AUTO] Fleet health probe stale",
body=_build_body(
fingerprint,
f"Detected at: {_iso(now)}",
"Heartbeat missing or older than the configured fleet health maximum age.",
),
)
)
return incidents
def find_matching_issue(incident: Incident, open_issues: Iterable[dict]) -> dict | None:
for issue in open_issues:
haystack = "\n".join([issue.get("title") or "", issue.get("body") or ""])
if incident.fingerprint in haystack or incident.title == issue.get("title"):
return issue
return None
def build_repeat_comment(incident: Incident) -> str:
return (
"Autonomous infrastructure detector saw the same incident again.\n\n"
f"Fingerprint: {incident.fingerprint}\n\n"
f"Latest evidence:\n{incident.latest_evidence()}"
)
def sync_incidents(
incidents: Iterable[Incident],
client: GiteaClient,
*,
apply: bool = False,
comment_existing: bool = True,
):
open_issues = list(client.list_open_issues())
results = []
for incident in incidents:
existing = find_matching_issue(incident, open_issues)
if existing:
action = "existing"
if apply and comment_existing:
client.comment_issue(existing["number"], build_repeat_comment(incident))
action = "commented"
results.append(
{
"action": action,
"fingerprint": incident.fingerprint,
"issue_number": existing["number"],
"title": existing.get("title"),
}
)
continue
if apply:
created = client.create_issue(incident.title, incident.body)
open_issues.append(created)
results.append(
{
"action": "created",
"fingerprint": incident.fingerprint,
"issue_number": created["number"],
"title": created.get("title"),
}
)
else:
results.append(
{
"action": "would_create",
"fingerprint": incident.fingerprint,
"issue_number": None,
"title": incident.title,
}
)
return results
def parse_args():
parser = argparse.ArgumentParser(description="Create or refresh fleet incidents on Gitea from local infrastructure signals.")
parser.add_argument("--owner", default=DEFAULT_OWNER)
parser.add_argument("--repo", default=DEFAULT_REPO)
parser.add_argument("--base-url", default=DEFAULT_BASE_URL)
parser.add_argument("--token-file", type=Path, default=DEFAULT_TOKEN_FILE)
parser.add_argument("--failover-status", type=Path, default=DEFAULT_FAILOVER_STATUS)
parser.add_argument("--restart-state-dir", type=Path, default=DEFAULT_RESTART_STATE_DIR)
parser.add_argument("--heartbeat-file", type=Path, default=DEFAULT_HEARTBEAT_FILE)
parser.add_argument("--heartbeat-max-age-seconds", type=int, default=900)
parser.add_argument("--restart-escalation-threshold", type=int, default=3)
parser.add_argument("--apply", action="store_true", help="Create/comment issues instead of reporting what would happen.")
parser.add_argument("--no-comment-existing", action="store_true", help="Do not comment on existing matching issues.")
parser.add_argument("--json", action="store_true", help="Emit machine-readable JSON output.")
return parser.parse_args()
def main():
args = parse_args()
now = datetime.now(timezone.utc)
failover_status = load_json(args.failover_status)
restart_counts = load_restart_counts(args.restart_state_dir)
heartbeat_stale = heartbeat_is_stale(
args.heartbeat_file,
now=now,
max_age_seconds=args.heartbeat_max_age_seconds,
)
incidents = build_incidents(
failover_status=failover_status,
restart_counts=restart_counts,
heartbeat_stale=heartbeat_stale,
now=now,
restart_escalation_threshold=args.restart_escalation_threshold,
)
payload = {
"generated_at": _iso(now),
"incidents": [incident.__dict__ for incident in incidents],
"results": [],
}
token = None
if args.token_file.exists():
token = args.token_file.read_text().strip()
if args.apply and not token:
raise SystemExit(f"Token file not found: {args.token_file}")
if token:
client = GiteaClient(token=token, owner=args.owner, repo=args.repo, base_url=args.base_url)
payload["results"] = sync_incidents(
incidents,
client,
apply=args.apply,
comment_existing=not args.no_comment_existing,
)
else:
payload["results"] = [
{
"action": "local_only",
"fingerprint": incident.fingerprint,
"issue_number": None,
"title": incident.title,
}
for incident in incidents
]
if args.json:
print(json.dumps(payload, indent=2))
else:
print(f"Generated at: {payload['generated_at']}")
if not incidents:
print("No autonomous infrastructure incidents detected.")
for incident in incidents:
print(f"- {incident.title} [{incident.fingerprint}]")
for result in payload["results"]:
print(f" -> {result['action']}: {result['title']}")
if __name__ == "__main__":
main()

110
scripts/backlog_cleanup.py Executable file
View File

@@ -0,0 +1,110 @@
#!/usr/bin/env python3
"""
Backlog Cleanup — Bulk close issues whose PRs are merged.
Usage:
python backlog_cleanup.py --repo Timmy_Foundation/timmy-home --dry-run
python backlog_cleanup.py --repo Timmy_Foundation/timmy-home --close
"""
import json
import os
import sys
import argparse
import urllib.request
import urllib.error
import time
from pathlib import Path
def get_token():
f = Path.home() / ".config" / "gitea" / "token"
if f.exists():
return f.read_text().strip()
return os.environ.get("GITEA_TOKEN", "")
def api(base, token, path, method="GET", data=None):
url = f"{base}/api/v1{path}"
headers = {"Authorization": f"token {token}"}
body = json.dumps(data).encode() if data else None
if data:
headers["Content-Type"] = "application/json"
req = urllib.request.Request(url, data=body, headers=headers, method=method)
try:
return json.loads(urllib.request.urlopen(req, timeout=15).read())
except Exception as e:
print(f" API error: {e}", file=sys.stderr)
return None
def main():
p = argparse.ArgumentParser()
p.add_argument("--repo", default="Timmy_Foundation/timmy-home")
p.add_argument("--base", default="https://forge.alexanderwhitestone.com")
p.add_argument("--dry-run", action="store_true", default=True)
p.add_argument("--close", action="store_true")
p.add_argument("--limit", type=int, default=20)
args = p.parse_args()
if args.close:
args.dry_run = False
token = get_token()
issues = api(args.base, token, f"/repos/{args.repo}/issues?state=open&limit={args.limit}")
if not issues:
return 1
issues = [i for i in issues if not i.get("pull_request")]
print(f"Scanning {len(issues)} issues...")
closable = []
for issue in issues:
if issue.get("assignees"):
continue
labels = {l.get("name", "").lower() for l in issue.get("labels", [])}
if labels & {"epic", "in-progress", "claw-code-in-progress", "blocked"}:
continue
# Check for merged PRs referencing this issue
ref = f"#{issue['number']}"
prs = api(args.base, token, f"/repos/{args.repo}/pulls?state=all&limit=20")
time.sleep(0.1) # Rate limit
linked_merged = [
pr for pr in (prs or [])
if ref in (pr.get("body", "") + pr.get("title", ""))
and (pr.get("state") == "merged" or pr.get("merged"))
]
if linked_merged:
reason = f"merged PR #{linked_merged[0]['number']}"
closable.append((issue, reason))
tag = "WOULD CLOSE" if args.dry_run else "CLOSING"
print(f" {tag} #{issue['number']}: {issue['title'][:50]}{reason}")
if not closable:
print("No issues to close.")
return 0
print(f"\n{'Would close' if args.dry_run else 'Closing'} {len(closable)} issues")
if args.dry_run:
print("(use --close to execute)")
return 0
closed = 0
for issue, reason in closable:
api(args.base, token, f"/repos/{args.repo}/issues/{issue['number']}/comments",
method="POST", data={"body": f"Closing — {reason}.\nAutomated by backlog_cleanup.py"})
r = api(args.base, token, f"/repos/{args.repo}/issues/{issue['number']}",
method="POST", data={"state": "closed"})
if r:
closed += 1
print(f" Closed #{issue['number']}")
time.sleep(0.2)
print(f"\nClosed {closed}/{len(closable)}")
return 0
if __name__ == "__main__":
sys.exit(main())

254
scripts/backlog_triage.py Executable file
View File

@@ -0,0 +1,254 @@
#!/usr/bin/env python3
"""
Weekly Backlog Triage for timmy-home
Issue #685: [OPS] timmy-home backlog reduced from 220 to 50 — triage cadence needed
Run this script weekly to maintain backlog visibility.
"""
import json
import os
import sys
import urllib.request
from datetime import datetime, timedelta
from typing import Any, Dict, List
# Configuration
GITEA_BASE = "https://forge.alexanderwhitestone.com/api/v1"
TOKEN_PATH = os.path.expanduser("~/.config/gitea/token")
ORG = "Timmy_Foundation"
REPO = "timmy-home"
class BacklogTriage:
"""Weekly backlog triage for timmy-home."""
def __init__(self):
self.token = self._load_token()
def _load_token(self) -> str:
"""Load Gitea API token."""
try:
with open(TOKEN_PATH, "r") as f:
return f.read().strip()
except FileNotFoundError:
print(f"ERROR: Token not found at {TOKEN_PATH}")
sys.exit(1)
def _api_request(self, endpoint: str) -> Any:
"""Make authenticated Gitea API request."""
url = f"{GITEA_BASE}{endpoint}"
headers = {"Authorization": f"token {self.token}"}
req = urllib.request.Request(url, headers=headers)
try:
with urllib.request.urlopen(req) as resp:
return json.loads(resp.read())
except urllib.error.HTTPError as e:
if e.code == 404:
return None
error_body = e.read().decode() if e.fp else "No error body"
print(f"API Error {e.code}: {error_body}")
return None
def get_open_issues(self) -> List[Dict]:
"""Get all open issues."""
endpoint = f"/repos/{ORG}/{REPO}/issues?state=open&limit=200"
issues = self._api_request(endpoint)
return issues if isinstance(issues, list) else []
def analyze_backlog(self, issues: List[Dict]) -> Dict[str, Any]:
"""Analyze the backlog."""
analysis = {
"total_open": len(issues),
"unassigned": 0,
"unlabeled": 0,
"batch_pipeline": 0,
"by_label": {},
"by_assignee": {},
"by_age": {
"0-7_days": 0,
"8-30_days": 0,
"31-90_days": 0,
"90+_days": 0
},
"stale_issues": [],
"unassigned_unlabeled": []
}
cutoff_date = datetime.now() - timedelta(days=30)
for issue in issues:
# Skip PRs
if 'pull_request' in issue:
continue
# Check assignment
if not issue.get('assignee'):
analysis["unassigned"] += 1
# Check labels
labels = [l['name'] for l in issue.get('labels', [])]
if not labels:
analysis["unlabeled"] += 1
else:
for label in labels:
analysis["by_label"][label] = analysis["by_label"].get(label, 0) + 1
# Check assignee
assignee = issue.get('assignee')
if assignee:
assignee_name = assignee['login']
analysis["by_assignee"][assignee_name] = analysis["by_assignee"].get(assignee_name, 0) + 1
# Check if batch-pipeline issue
if 'batch-pipeline' in labels:
analysis["batch_pipeline"] += 1
# Check age
created_at = datetime.fromisoformat(issue['created_at'].replace('Z', '+00:00'))
age_days = (datetime.now() - created_at).days
if age_days <= 7:
analysis["by_age"]["0-7_days"] += 1
elif age_days <= 30:
analysis["by_age"]["8-30_days"] += 1
elif age_days <= 90:
analysis["by_age"]["31-90_days"] += 1
else:
analysis["by_age"]["90+_days"] += 1
# Check if stale (>30 days old and no labels/assignee)
if age_days > 30 and not labels and not issue.get('assignee'):
analysis["stale_issues"].append({
"number": issue['number'],
"title": issue['title'],
"created": issue['created_at'],
"age_days": age_days
})
# Check if unassigned and unlabeled
if not issue.get('assignee') and not labels:
analysis["unassigned_unlabeled"].append({
"number": issue['number'],
"title": issue['title'],
"created": issue['created_at']
})
return analysis
def generate_report(self, analysis: Dict[str, Any]) -> str:
"""Generate a triage report."""
report = f"# timmy-home Weekly Backlog Triage\n\n"
report += f"Generated: {datetime.now().isoformat()}\n\n"
report += "## Summary\n"
report += f"- **Total open issues:** {analysis['total_open']}\n"
report += f"- **Unassigned:** {analysis['unassigned']}\n"
report += f"- **Unlabeled:** {analysis['unlabeled']}\n"
report += f"- **Batch-pipeline issues:** {analysis['batch_pipeline']}\n"
report += f"- **Stale issues (>30 days, no labels/assignee):** {len(analysis['stale_issues'])}\n"
report += f"- **Unassigned + Unlabeled:** {len(analysis['unassigned_unlabeled'])}\n\n"
report += "## Age Distribution\n"
for age_range, count in analysis['by_age'].items():
report += f"- **{age_range}:** {count} issues\n"
report += "\n## Label Distribution\n"
if analysis['by_label']:
for label, count in sorted(analysis['by_label'].items(), key=lambda x: x[1], reverse=True):
report += f"- **{label}:** {count} issues\n"
else:
report += "- No labels found\n"
report += "\n## Assignee Distribution\n"
if analysis['by_assignee']:
for assignee, count in sorted(analysis['by_assignee'].items(), key=lambda x: x[1], reverse=True):
report += f"- **@{assignee}:** {count} issues\n"
else:
report += "- No assignees found\n"
if analysis['stale_issues']:
report += "\n## Stale Issues (>30 days, no labels/assignee)\n"
report += "These issues should be triaged or closed:\n"
for issue in analysis['stale_issues'][:10]: # Show first 10
report += f"- **#{issue['number']}**: {issue['title']}\n"
report += f" - Age: {issue['age_days']} days\n"
report += f" - Created: {issue['created']}\n"
if analysis['unassigned_unlabeled']:
report += "\n## Unassigned + Unlabeled Issues\n"
report += "These issues need labels and/or assignees:\n"
for issue in analysis['unassigned_unlabeled'][:10]: # Show first 10
report += f"- **#{issue['number']}**: {issue['title']}\n"
report += f" - Created: {issue['created']}\n"
report += "\n## Recommendations\n"
if analysis['unassigned'] > 0:
report += f"1. **Assign owners to {analysis['unassigned']} issues** - Ensure accountability\n"
if analysis['unlabeled'] > 0:
report += f"2. **Add labels to {analysis['unlabeled']} issues** - Categorize for management\n"
if len(analysis['stale_issues']) > 0:
report += f"3. **Triage {len(analysis['stale_issues'])} stale issues** - Close or re-prioritize\n"
if len(analysis['unassigned_unlabeled']) > 0:
report += f"4. **Address {len(analysis['unassigned_unlabeled'])} unassigned/unlabeled issues** - Basic triage needed\n"
return report
def generate_cron_entry(self) -> str:
"""Generate cron entry for weekly triage."""
cron_entry = """# Weekly timmy-home backlog triage
# Run every Monday at 9:00 AM
0 9 * * 1 cd /path/to/timmy-home && python3 scripts/backlog_triage.py --report > /var/log/timmy-home-triage-$(date +\\%Y\\%m\\%d).log 2>&1
# Or run directly:
# python3 scripts/backlog_triage.py --report"""
return cron_entry
def main():
"""Main entry point."""
import argparse
parser = argparse.ArgumentParser(description="Weekly Backlog Triage for timmy-home")
parser.add_argument("--analyze", action="store_true", help="Analyze backlog")
parser.add_argument("--report", action="store_true", help="Generate report")
parser.add_argument("--cron", action="store_true", help="Generate cron entry")
parser.add_argument("--json", action="store_true", help="Output JSON")
args = parser.parse_args()
triage = BacklogTriage()
if args.analyze or args.report or args.json:
issues = triage.get_open_issues()
analysis = triage.analyze_backlog(issues)
if args.json:
print(json.dumps(analysis, indent=2))
elif args.report:
report = triage.generate_report(analysis)
print(report)
else:
# Default: show summary
print(f"timmy-home Backlog Analysis:")
print(f" Total open issues: {analysis['total_open']}")
print(f" Unassigned: {analysis['unassigned']}")
print(f" Unlabeled: {analysis['unlabeled']}")
print(f" Batch-pipeline: {analysis['batch_pipeline']}")
print(f" Stale issues: {len(analysis['stale_issues'])}")
print(f" Unassigned + Unlabeled: {len(analysis['unassigned_unlabeled'])}")
elif args.cron:
# Generate cron entry
cron_entry = triage.generate_cron_entry()
print(cron_entry)
else:
parser.print_help()
if __name__ == "__main__":
main()

22
scripts/backlog_triage_cron.sh Executable file
View File

@@ -0,0 +1,22 @@
#!/usr/bin/env bash
# backlog_triage_cron.sh — Weekly cron wrapper for backlog_triage.py
# Add to crontab: 0 9 * * 1 /path/to/timmy-home/scripts/backlog_triage_cron.sh
# Runs Monday 9am UTC
set -euo pipefail
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
REPO_DIR="$(dirname "$SCRIPT_DIR")"
REPORT_DIR="$REPO_DIR/reports/production"
REPORT_FILE="$REPORT_DIR/backlog_triage_$(date +%Y%m%d).md"
mkdir -p "$REPORT_DIR"
# Run triage, capture output
OUTPUT=$("$SCRIPT_DIR/backlog_triage.py" 2>&1) || true
# Save report
echo "$OUTPUT" > "$REPORT_FILE"
# Print to stdout for cron logging
echo "$OUTPUT"

View File

@@ -1,80 +1,170 @@
#!/usr/bin/env bash
# backup_pipeline.sh — Daily fleet backup pipeline (FLEET-008)
# Refs: timmy-home #561
# backup_pipeline.sh — Nightly encrypted Hermes backup pipeline
# Refs: timmy-home #693, timmy-home #561
set -euo pipefail
BACKUP_ROOT="/backups/timmy"
DATESTAMP=$(date +%Y%m%d-%H%M%S)
BACKUP_DIR="${BACKUP_ROOT}/${DATESTAMP}"
LOG_DIR="/var/log/timmy"
ALERT_LOG="${LOG_DIR}/backup_pipeline.log"
mkdir -p "$BACKUP_DIR" "$LOG_DIR"
DATESTAMP="${BACKUP_TIMESTAMP:-$(date +%Y%m%d-%H%M%S)}"
BACKUP_SOURCE_DIR="${BACKUP_SOURCE_DIR:-${HOME}/.hermes}"
BACKUP_ROOT="${BACKUP_ROOT:-${HOME}/.timmy-backups/hermes}"
BACKUP_LOG_DIR="${BACKUP_LOG_DIR:-${BACKUP_ROOT}/logs}"
BACKUP_RETENTION_DAYS="${BACKUP_RETENTION_DAYS:-14}"
BACKUP_S3_URI="${BACKUP_S3_URI:-}"
BACKUP_NAS_TARGET="${BACKUP_NAS_TARGET:-}"
AWS_ENDPOINT_URL="${AWS_ENDPOINT_URL:-}"
BACKUP_NAME="hermes-backup-${DATESTAMP}"
LOCAL_BACKUP_DIR="${BACKUP_ROOT}/${DATESTAMP}"
STAGE_DIR="$(mktemp -d "${TMPDIR:-/tmp}/timmy-backup.XXXXXX")"
PLAINTEXT_ARCHIVE="${STAGE_DIR}/${BACKUP_NAME}.tar.gz"
ENCRYPTED_ARCHIVE="${STAGE_DIR}/${BACKUP_NAME}.tar.gz.enc"
MANIFEST_PATH="${STAGE_DIR}/${BACKUP_NAME}.json"
ALERT_LOG="${BACKUP_LOG_DIR}/backup_pipeline.log"
PASSFILE_CLEANUP=""
TELEGRAM_BOT_TOKEN="${TELEGRAM_BOT_TOKEN:-}"
TELEGRAM_CHAT_ID="${TELEGRAM_CHAT_ID:-}"
OFFSITE_TARGET="${OFFSITE_TARGET:-}"
mkdir -p "$BACKUP_LOG_DIR"
log() { echo "[$(date -Iseconds)] $1" | tee -a "$ALERT_LOG"; }
log() {
echo "[$(date -Iseconds)] $1" | tee -a "$ALERT_LOG"
}
send_telegram() {
local msg="$1"
if [[ -n "$TELEGRAM_BOT_TOKEN" && -n "$TELEGRAM_CHAT_ID" ]]; then
curl -s -X POST "https://api.telegram.org/bot${TELEGRAM_BOT_TOKEN}/sendMessage" \
-d "chat_id=${TELEGRAM_CHAT_ID}" -d "text=${msg}" >/dev/null 2>&1 || true
fail() {
log "ERROR: $1"
exit 1
}
cleanup() {
rm -f "$PLAINTEXT_ARCHIVE"
rm -rf "$STAGE_DIR"
if [[ -n "$PASSFILE_CLEANUP" && -f "$PASSFILE_CLEANUP" ]]; then
rm -f "$PASSFILE_CLEANUP"
fi
}
trap cleanup EXIT
resolve_passphrase_file() {
if [[ -n "${BACKUP_PASSPHRASE_FILE:-}" ]]; then
[[ -f "$BACKUP_PASSPHRASE_FILE" ]] || fail "BACKUP_PASSPHRASE_FILE does not exist: $BACKUP_PASSPHRASE_FILE"
echo "$BACKUP_PASSPHRASE_FILE"
return
fi
if [[ -n "${BACKUP_PASSPHRASE:-}" ]]; then
PASSFILE_CLEANUP="${STAGE_DIR}/backup.passphrase"
printf '%s' "$BACKUP_PASSPHRASE" > "$PASSFILE_CLEANUP"
chmod 600 "$PASSFILE_CLEANUP"
echo "$PASSFILE_CLEANUP"
return
fi
fail "Set BACKUP_PASSPHRASE_FILE or BACKUP_PASSPHRASE before running the backup pipeline."
}
sha256_file() {
local path="$1"
if command -v shasum >/dev/null 2>&1; then
shasum -a 256 "$path" | awk '{print $1}'
elif command -v sha256sum >/dev/null 2>&1; then
sha256sum "$path" | awk '{print $1}'
else
python3 - <<'PY' "$path"
import hashlib
import pathlib
import sys
path = pathlib.Path(sys.argv[1])
h = hashlib.sha256()
with path.open('rb') as f:
for chunk in iter(lambda: f.read(1024 * 1024), b''):
h.update(chunk)
print(h.hexdigest())
PY
fi
}
status=0
write_manifest() {
python3 - <<'PY' "$1" "$2" "$3" "$4" "$5" "$6" "$7" "$8"
import json
import sys
manifest_path, source_dir, archive_name, archive_sha256, local_dir, s3_uri, nas_target, created_at = sys.argv[1:]
manifest = {
"created_at": created_at,
"source_dir": source_dir,
"archive_name": archive_name,
"archive_sha256": archive_sha256,
"encryption": {
"type": "openssl",
"cipher": "aes-256-cbc",
"pbkdf2": True,
"iterations": 200000,
},
"destinations": {
"local_dir": local_dir,
"s3_uri": s3_uri or None,
"nas_target": nas_target or None,
},
}
with open(manifest_path, 'w', encoding='utf-8') as handle:
json.dump(manifest, handle, indent=2)
handle.write('\n')
PY
}
# --- Gitea repositories ---
if [[ -d /root/gitea ]]; then
tar czf "${BACKUP_DIR}/gitea-repos.tar.gz" -C /root gitea 2>/dev/null || true
log "Backed up Gitea repos"
fi
upload_to_nas() {
local archive_path="$1"
local manifest_path="$2"
local target_root="$3"
# --- Agent configs and state ---
for wiz in bezalel allegro ezra timmy; do
if [[ -d "/root/wizards/${wiz}" ]]; then
tar czf "${BACKUP_DIR}/${wiz}-home.tar.gz" -C /root/wizards "${wiz}" 2>/dev/null || true
log "Backed up ${wiz} home"
local target_dir="${target_root%/}/${DATESTAMP}"
mkdir -p "$target_dir"
cp "$archive_path" "$manifest_path" "$target_dir/"
log "Uploaded backup to NAS target: $target_dir"
}
upload_to_s3() {
local archive_path="$1"
local manifest_path="$2"
command -v aws >/dev/null 2>&1 || fail "BACKUP_S3_URI is set but aws CLI is not installed."
local args=()
if [[ -n "$AWS_ENDPOINT_URL" ]]; then
args+=(--endpoint-url "$AWS_ENDPOINT_URL")
fi
done
# --- System configs ---
cp /etc/crontab "${BACKUP_DIR}/crontab" 2>/dev/null || true
cp -r /etc/systemd/system "${BACKUP_DIR}/systemd" 2>/dev/null || true
log "Backed up system configs"
aws "${args[@]}" s3 cp "$archive_path" "${BACKUP_S3_URI%/}/$(basename "$archive_path")"
aws "${args[@]}" s3 cp "$manifest_path" "${BACKUP_S3_URI%/}/$(basename "$manifest_path")"
log "Uploaded backup to S3 target: $BACKUP_S3_URI"
}
# --- Evennia worlds (if present) ---
if [[ -d /root/evennia ]]; then
tar czf "${BACKUP_DIR}/evennia-worlds.tar.gz" -C /root evennia 2>/dev/null || true
log "Backed up Evennia worlds"
[[ -d "$BACKUP_SOURCE_DIR" ]] || fail "BACKUP_SOURCE_DIR does not exist: $BACKUP_SOURCE_DIR"
[[ -n "$BACKUP_NAS_TARGET" || -n "$BACKUP_S3_URI" ]] || fail "Set BACKUP_NAS_TARGET or BACKUP_S3_URI for remote backup storage."
PASSFILE="$(resolve_passphrase_file)"
mkdir -p "$LOCAL_BACKUP_DIR"
log "Creating archive from $BACKUP_SOURCE_DIR"
tar -czf "$PLAINTEXT_ARCHIVE" -C "$(dirname "$BACKUP_SOURCE_DIR")" "$(basename "$BACKUP_SOURCE_DIR")"
log "Encrypting archive"
openssl enc -aes-256-cbc -salt -pbkdf2 -iter 200000 \
-pass "file:${PASSFILE}" \
-in "$PLAINTEXT_ARCHIVE" \
-out "$ENCRYPTED_ARCHIVE"
ARCHIVE_SHA256="$(sha256_file "$ENCRYPTED_ARCHIVE")"
CREATED_AT="$(date -u '+%Y-%m-%dT%H:%M:%SZ')"
write_manifest "$MANIFEST_PATH" "$BACKUP_SOURCE_DIR" "$(basename "$ENCRYPTED_ARCHIVE")" "$ARCHIVE_SHA256" "$LOCAL_BACKUP_DIR" "$BACKUP_S3_URI" "$BACKUP_NAS_TARGET" "$CREATED_AT"
cp "$ENCRYPTED_ARCHIVE" "$MANIFEST_PATH" "$LOCAL_BACKUP_DIR/"
rm -f "$PLAINTEXT_ARCHIVE"
log "Encrypted backup stored locally: ${LOCAL_BACKUP_DIR}/$(basename "$ENCRYPTED_ARCHIVE")"
if [[ -n "$BACKUP_NAS_TARGET" ]]; then
upload_to_nas "$ENCRYPTED_ARCHIVE" "$MANIFEST_PATH" "$BACKUP_NAS_TARGET"
fi
# --- Manifest ---
find "$BACKUP_DIR" -type f > "${BACKUP_DIR}/manifest.txt"
log "Backup manifest written"
# --- Offsite sync ---
if [[ -n "$OFFSITE_TARGET" ]]; then
if rsync -az --delete "${BACKUP_DIR}/" "${OFFSITE_TARGET}/${DATESTAMP}/" 2>/dev/null; then
log "Offsite sync completed"
else
log "WARNING: Offsite sync failed"
status=1
fi
if [[ -n "$BACKUP_S3_URI" ]]; then
upload_to_s3 "$ENCRYPTED_ARCHIVE" "$MANIFEST_PATH"
fi
# --- Retention: keep last 7 days ---
find "$BACKUP_ROOT" -mindepth 1 -maxdepth 1 -type d -mtime +7 -exec rm -rf {} + 2>/dev/null || true
log "Retention applied (7 days)"
if [[ "$status" -eq 0 ]]; then
log "Backup pipeline completed: ${BACKUP_DIR}"
send_telegram "✅ Daily backup completed: ${DATESTAMP}"
else
log "Backup pipeline completed with WARNINGS: ${BACKUP_DIR}"
send_telegram "⚠️ Daily backup completed with warnings: ${DATESTAMP}"
fi
exit "$status"
find "$BACKUP_ROOT" -mindepth 1 -maxdepth 1 -type d -name '20*' -mtime "+${BACKUP_RETENTION_DAYS}" -exec rm -rf {} + 2>/dev/null || true
log "Retention applied (${BACKUP_RETENTION_DAYS} days)"
log "Backup pipeline completed successfully"

View File

@@ -0,0 +1,228 @@
#!/usr/bin/env python3
"""Provisioning and wiring scaffold for Bezalel Gemma 4 on RunPod.
Refs: timmy-home #544
Safe by default:
- builds the RunPod deploy mutation
- can call the RunPod GraphQL API if a key is provided and --apply-runpod is used
- can update a Hermes config file in-place when --write-config is used
- can verify an OpenAI-compatible endpoint with a lightweight chat probe
"""
from __future__ import annotations
import argparse
import json
from pathlib import Path
from typing import Any
from urllib import request
import yaml
RUNPOD_GRAPHQL_URL = "https://api.runpod.io/graphql"
DEFAULT_GPU_TYPE = "NVIDIA L40S"
DEFAULT_CLOUD_TYPE = "COMMUNITY"
DEFAULT_IMAGE = "ollama/ollama:latest"
DEFAULT_MODEL = "gemma4:latest"
DEFAULT_PROVIDER_NAME = "Big Brain"
DEFAULT_TOKEN_FILE = Path.home() / ".config" / "runpod" / "access_key"
DEFAULT_CONFIG_PATH = Path.home() / "wizards" / "bezalel" / "home" / "config.yaml"
def build_deploy_mutation(
*,
name: str,
gpu_type: str = DEFAULT_GPU_TYPE,
cloud_type: str = DEFAULT_CLOUD_TYPE,
container_disk_gb: int = 100,
volume_gb: int = 50,
model_tag: str = DEFAULT_MODEL,
) -> str:
# model_tag is accepted for parity with the CLI/reporting path even though the
# pod deploy itself only needs the Ollama image + port wiring.
_ = model_tag
return f'''
mutation {{
podFindAndDeployOnDemand(input: {{
cloudType: {cloud_type},
gpuCount: 1,
gpuTypeId: "{gpu_type}",
name: "{name}",
containerDiskInGb: {container_disk_gb},
imageName: "{DEFAULT_IMAGE}",
ports: "11434/http",
volumeInGb: {volume_gb},
volumeMountPath: "/root/.ollama"
}}) {{
id
desiredStatus
machineId
}}
}}
'''.strip()
def build_runpod_endpoint(pod_id: str, port: int = 11434) -> str:
return f"https://{pod_id}-{port}.proxy.runpod.net/v1"
def parse_deploy_response(payload: dict[str, Any]) -> dict[str, str]:
data = (payload.get("data") or {}).get("podFindAndDeployOnDemand") or {}
pod_id = data.get("id")
if not pod_id:
raise ValueError(f"RunPod deploy response did not contain a pod id: {payload}")
return {
"pod_id": pod_id,
"desired_status": data.get("desiredStatus", "UNKNOWN"),
"base_url": build_runpod_endpoint(pod_id),
}
def deploy_runpod(*, api_key: str, name: str, gpu_type: str = DEFAULT_GPU_TYPE, cloud_type: str = DEFAULT_CLOUD_TYPE, model: str = DEFAULT_MODEL) -> dict[str, str]:
query = build_deploy_mutation(name=name, gpu_type=gpu_type, cloud_type=cloud_type, model_tag=model)
payload = json.dumps({"query": query}).encode()
req = request.Request(
RUNPOD_GRAPHQL_URL,
data=payload,
headers={
"Authorization": f"Bearer {api_key}",
"Content-Type": "application/json",
},
method="POST",
)
with request.urlopen(req, timeout=30) as resp:
response_payload = json.loads(resp.read().decode())
return parse_deploy_response(response_payload)
def update_config_text(config_text: str, *, base_url: str, model: str = DEFAULT_MODEL, provider_name: str = DEFAULT_PROVIDER_NAME) -> str:
parsed = yaml.safe_load(config_text) or {}
providers = list(parsed.get("custom_providers") or [])
replacement = {
"name": provider_name,
"base_url": base_url,
"api_key": "",
"model": model,
}
updated = False
for idx, provider in enumerate(providers):
if provider.get("name") == provider_name:
providers[idx] = replacement
updated = True
break
if not updated:
providers.append(replacement)
parsed["custom_providers"] = providers
return yaml.safe_dump(parsed, sort_keys=False)
def write_config_file(config_path: Path, *, base_url: str, model: str = DEFAULT_MODEL, provider_name: str = DEFAULT_PROVIDER_NAME) -> str:
original = config_path.read_text() if config_path.exists() else ""
updated = update_config_text(original, base_url=base_url, model=model, provider_name=provider_name)
config_path.parent.mkdir(parents=True, exist_ok=True)
config_path.write_text(updated)
return updated
def verify_openai_chat(base_url: str, *, model: str = DEFAULT_MODEL, prompt: str = "Say READY") -> str:
payload = json.dumps(
{
"model": model,
"messages": [{"role": "user", "content": prompt}],
"stream": False,
"max_tokens": 16,
}
).encode()
req = request.Request(
f"{base_url.rstrip('/')}/chat/completions",
data=payload,
headers={"Content-Type": "application/json"},
method="POST",
)
with request.urlopen(req, timeout=30) as resp:
data = json.loads(resp.read().decode())
return data["choices"][0]["message"]["content"]
def parse_args() -> argparse.Namespace:
parser = argparse.ArgumentParser(description="Provision a RunPod Gemma 4 endpoint and wire a Hermes config for Bezalel.")
parser.add_argument("--pod-name", default="bezalel-gemma4")
parser.add_argument("--gpu-type", default=DEFAULT_GPU_TYPE)
parser.add_argument("--cloud-type", default=DEFAULT_CLOUD_TYPE)
parser.add_argument("--model", default=DEFAULT_MODEL)
parser.add_argument("--provider-name", default=DEFAULT_PROVIDER_NAME)
parser.add_argument("--token-file", type=Path, default=DEFAULT_TOKEN_FILE)
parser.add_argument("--config-path", type=Path, default=DEFAULT_CONFIG_PATH)
parser.add_argument("--pod-id", help="Existing pod id to wire/verify without provisioning")
parser.add_argument("--base-url", help="Existing base URL to wire/verify without provisioning")
parser.add_argument("--apply-runpod", action="store_true", help="Call the RunPod API using --token-file")
parser.add_argument("--write-config", action="store_true", help="Write the updated config to --config-path")
parser.add_argument("--verify-chat", action="store_true", help="Call the OpenAI-compatible chat endpoint")
parser.add_argument("--json", action="store_true", help="Emit machine-readable JSON")
return parser.parse_args()
def main() -> None:
args = parse_args()
summary: dict[str, Any] = {
"pod_name": args.pod_name,
"gpu_type": args.gpu_type,
"cloud_type": args.cloud_type,
"model": args.model,
"provider_name": args.provider_name,
"actions": [],
}
base_url = args.base_url
if not base_url and args.pod_id:
base_url = build_runpod_endpoint(args.pod_id)
summary["actions"].append("computed_base_url_from_pod_id")
if args.apply_runpod:
if not args.token_file.exists():
raise SystemExit(f"RunPod token file not found: {args.token_file}")
api_key = args.token_file.read_text().strip()
deployed = deploy_runpod(api_key=api_key, name=args.pod_name, gpu_type=args.gpu_type, cloud_type=args.cloud_type, model=args.model)
summary["deployment"] = deployed
base_url = deployed["base_url"]
summary["actions"].append("deployed_runpod_pod")
if not base_url:
base_url = build_runpod_endpoint("<pod-id>")
summary["actions"].append("using_placeholder_base_url")
summary["base_url"] = base_url
summary["config_preview"] = update_config_text("", base_url=base_url, model=args.model, provider_name=args.provider_name)
if args.write_config:
write_config_file(args.config_path, base_url=base_url, model=args.model, provider_name=args.provider_name)
summary["config_path"] = str(args.config_path)
summary["actions"].append("wrote_config")
if args.verify_chat:
summary["verify_response"] = verify_openai_chat(base_url, model=args.model)
summary["actions"].append("verified_chat")
if args.json:
print(json.dumps(summary, indent=2))
return
print("--- Bezalel Gemma4 RunPod Wiring ---")
print(f"Pod name: {args.pod_name}")
print(f"Base URL: {base_url}")
print(f"Model: {args.model}")
if args.write_config:
print(f"Config written: {args.config_path}")
if "verify_response" in summary:
print(f"Verify response: {summary['verify_response']}")
if summary["actions"]:
print("Actions: " + ", ".join(summary["actions"]))
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,158 @@
#!/usr/bin/env python3
"""Bezalel Tailscale bootstrap scaffold.
Refs: timmy-home #535
Safe by default:
- builds a remote bootstrap shell script
- can write that script to disk
- can print the SSH command needed to execute it
- only runs remote SSH when --apply is explicitly passed
"""
from __future__ import annotations
import argparse
import json
import shlex
import subprocess
from pathlib import Path
from typing import Any
DEFAULT_HOST = "159.203.146.185"
DEFAULT_HOSTNAME = "bezalel"
DEFAULT_PEERS = {
"mac": "100.124.176.28",
"ezra": "100.126.61.75",
}
def build_remote_script(
*,
auth_key: str,
ssh_public_key: str,
peers: dict[str, str] | None = None,
hostname: str = DEFAULT_HOSTNAME,
) -> str:
peer_map = peers or DEFAULT_PEERS
lines = [
"#!/usr/bin/env bash",
"set -euo pipefail",
"curl -fsSL https://tailscale.com/install.sh | sh",
f"tailscale up --authkey {shlex.quote(auth_key)} --ssh --hostname {shlex.quote(hostname)}",
"install -d -m 700 ~/.ssh",
f"touch ~/.ssh/authorized_keys && chmod 600 ~/.ssh/authorized_keys",
f"grep -qxF {shlex.quote(ssh_public_key)} ~/.ssh/authorized_keys || printf '%s\\n' {shlex.quote(ssh_public_key)} >> ~/.ssh/authorized_keys",
"tailscale status --json",
]
for name, ip in peer_map.items():
lines.append(f"ping -c 1 {shlex.quote(ip)} >/dev/null && echo 'PING_OK:{name}:{ip}'")
return "\n".join(lines) + "\n"
def parse_tailscale_status(payload: dict[str, Any]) -> dict[str, Any]:
self_block = payload.get("Self") or {}
peers = payload.get("Peer") or {}
return {
"self": {
"hostname": self_block.get("HostName"),
"dns_name": self_block.get("DNSName"),
"tailscale_ips": list(self_block.get("TailscaleIPs") or []),
},
"peers": {
peer.get("HostName") or peer_key: list(peer.get("TailscaleIPs") or [])
for peer_key, peer in peers.items()
},
}
def build_ssh_command(host: str, remote_script_path: str = "/tmp/bezalel_tailscale_bootstrap.sh") -> list[str]:
return ["ssh", host, f"bash {shlex.quote(remote_script_path)}"]
def write_script(path: Path, content: str) -> None:
path.parent.mkdir(parents=True, exist_ok=True)
path.write_text(content)
def run_remote(host: str, remote_script_path: str) -> subprocess.CompletedProcess[str]:
return subprocess.run(build_ssh_command(host, remote_script_path), capture_output=True, text=True, timeout=120)
def parse_peer_args(items: list[str]) -> dict[str, str]:
peers = dict(DEFAULT_PEERS)
for item in items:
name, ip = item.split("=", 1)
peers[name.strip()] = ip.strip()
return peers
def parse_args() -> argparse.Namespace:
parser = argparse.ArgumentParser(description="Prepare or execute Tailscale bootstrap for the Bezalel VPS.")
parser.add_argument("--host", default=DEFAULT_HOST)
parser.add_argument("--hostname", default=DEFAULT_HOSTNAME)
parser.add_argument("--auth-key", help="Tailscale auth key")
parser.add_argument("--auth-key-file", type=Path, help="Path to file containing the Tailscale auth key")
parser.add_argument("--ssh-public-key", help="SSH public key to append to authorized_keys")
parser.add_argument("--ssh-public-key-file", type=Path, help="Path to the SSH public key file")
parser.add_argument("--peer", action="append", default=[], help="Additional peer as name=ip")
parser.add_argument("--script-out", type=Path, default=Path("/tmp/bezalel_tailscale_bootstrap.sh"))
parser.add_argument("--remote-script-path", default="/tmp/bezalel_tailscale_bootstrap.sh")
parser.add_argument("--apply", action="store_true", help="Execute the generated script over SSH")
parser.add_argument("--json", action="store_true")
return parser.parse_args()
def _read_secret(value: str | None, path: Path | None) -> str | None:
if value:
return value.strip()
if path and path.exists():
return path.read_text().strip()
return None
def main() -> None:
args = parse_args()
auth_key = _read_secret(args.auth_key, args.auth_key_file)
ssh_public_key = _read_secret(args.ssh_public_key, args.ssh_public_key_file)
peers = parse_peer_args(args.peer)
if not auth_key:
raise SystemExit("Missing Tailscale auth key. Use --auth-key or --auth-key-file.")
if not ssh_public_key:
raise SystemExit("Missing SSH public key. Use --ssh-public-key or --ssh-public-key-file.")
script = build_remote_script(auth_key=auth_key, ssh_public_key=ssh_public_key, peers=peers, hostname=args.hostname)
write_script(args.script_out, script)
payload: dict[str, Any] = {
"host": args.host,
"hostname": args.hostname,
"script_out": str(args.script_out),
"remote_script_path": args.remote_script_path,
"ssh_command": build_ssh_command(args.host, args.remote_script_path),
"peer_targets": peers,
"applied": False,
}
if args.apply:
result = run_remote(args.host, args.remote_script_path)
payload["applied"] = True
payload["exit_code"] = result.returncode
payload["stdout"] = result.stdout
payload["stderr"] = result.stderr
if args.json:
print(json.dumps(payload, indent=2))
return
print("--- Bezalel Tailscale Bootstrap ---")
print(f"Host: {args.host}")
print(f"Local script: {args.script_out}")
print("SSH command: " + " ".join(payload["ssh_command"]))
if args.apply:
print(f"Exit code: {payload['exit_code']}")
if __name__ == "__main__":
main()

View File

@@ -1,214 +1,123 @@
#!/usr/bin/env python3
"""
Big Brain Pod Management and Verification
Comprehensive script for managing and verifying Big Brain pod.
Big Brain provider management and verification.
Uses the repo's Big Brain provider config rather than a stale hardcoded pod id.
Supports both OpenAI-compatible and raw Ollama backends.
"""
import requests
import time
from __future__ import annotations
import json
import os
import sys
from datetime import datetime
# Configuration
CONFIG = {
"pod_id": "8lfr3j47a5r3gn",
"endpoint": "https://8lfr3j47a5r3gn-11434.proxy.runpod.net",
"cost_per_hour": 0.79,
"model": "gemma3:27b",
"max_response_time": 30, # seconds
"timeout": 10
}
import requests
class PodVerifier:
def __init__(self, config=None):
self.config = config or CONFIG
self.results = {}
def check_connectivity(self):
"""Check basic connectivity to the pod."""
print(f"[{datetime.now().isoformat()}] Checking connectivity to {self.config['endpoint']}...")
from scripts.big_brain_provider import (
build_generate_payload,
resolve_big_brain_provider,
resolve_generate_url,
resolve_models_url,
)
class ProviderVerifier:
def __init__(self, provider: dict | None = None, timeout: int = 10, max_response_time: int = 30):
self.provider = provider or resolve_big_brain_provider()
self.timeout = timeout
self.max_response_time = max_response_time
self.results: dict[str, object] = {}
def _headers(self) -> dict[str, str]:
headers = {"Content-Type": "application/json"}
api_key = self.provider.get("api_key", "")
if api_key:
headers["Authorization"] = f"Bearer {api_key}"
return headers
def check_models(self):
url = resolve_models_url(self.provider)
print(f"[{datetime.now().isoformat()}] Checking models endpoint: {url}")
try:
response = requests.get(self.config['endpoint'], timeout=self.config['timeout'])
print(f" Status: {response.status_code}")
print(f" Headers: {dict(response.headers)}")
return response.status_code
except requests.exceptions.ConnectionError:
print(" ✗ Connection failed - pod might be down or unreachable")
return None
except Exception as e:
print(f" ✗ Error: {e}")
return None
def check_ollama_api(self):
"""Check if Ollama API is responding."""
print(f"[{datetime.now().isoformat()}] Checking Ollama API...")
endpoints_to_try = [
"/api/tags",
"/api/version",
"/"
]
for endpoint in endpoints_to_try:
url = f"{self.config['endpoint']}{endpoint}"
try:
print(f" Trying {url}...")
response = requests.get(url, timeout=self.config['timeout'])
print(f" Status: {response.status_code}")
if response.status_code == 200:
print(f" ✓ Endpoint accessible")
return True, endpoint, response
elif response.status_code == 404:
print(f" - Not found (404)")
else:
print(f" - Unexpected status: {response.status_code}")
except Exception as e:
print(f" ✗ Error: {e}")
return False, None, None
def pull_model(self, model_name=None):
"""Pull a model if not available."""
model = model_name or self.config['model']
print(f"[{datetime.now().isoformat()}] Pulling model {model}...")
try:
payload = {"name": model}
response = requests.post(
f"{self.config['endpoint']}/api/pull",
json=payload,
timeout=60
)
if response.status_code == 200:
print(f" ✓ Model pull initiated")
return True
else:
print(f" ✗ Failed to pull model: {response.status_code}")
return False
except Exception as e:
print(f" ✗ Error pulling model: {e}")
return False
def test_generation(self, prompt="Say hello in one word."):
"""Test generation with the model."""
print(f"[{datetime.now().isoformat()}] Testing generation...")
try:
payload = {
"model": self.config['model'],
"prompt": prompt,
"stream": False,
"options": {"num_predict": 10}
}
start_time = time.time()
response = requests.post(
f"{self.config['endpoint']}/api/generate",
json=payload,
timeout=self.config['max_response_time']
)
elapsed = time.time() - start_time
response = requests.get(url, headers=self._headers(), timeout=self.timeout)
models = []
if response.status_code == 200:
data = response.json()
response_text = data.get("response", "").strip()
print(f" ✓ Generation successful in {elapsed:.2f}s")
print(f" Response: {response_text[:100]}...")
if elapsed <= self.config['max_response_time']:
print(f" ✓ Response time within limit ({self.config['max_response_time']}s)")
return True, elapsed, response_text
if self.provider["backend"] == "openai":
models = [m.get("id", "") for m in data.get("data", [])]
else:
print(f" ✗ Response time {elapsed:.2f}s exceeds limit")
return False, elapsed, response_text
models = [m.get("name", "") for m in data.get("models", [])]
print(f" ✓ Models endpoint OK ({response.status_code})")
else:
print(f"Generation failed: {response.status_code}")
return False, 0, ""
print(f"Models endpoint failed ({response.status_code})")
return response.status_code == 200, models, response.status_code
except Exception as e:
print(f"Error during generation: {e}")
return False, 0, ""
def run_verification(self):
"""Run full verification suite."""
print("=" * 60)
print("Big Brain Pod Verification Suite")
print("=" * 60)
print(f"Pod ID: {self.config['pod_id']}")
print(f"Endpoint: {self.config['endpoint']}")
print(f"Model: {self.config['model']}")
print(f"Cost: ${self.config['cost_per_hour']}/hour")
print("=" * 60)
print()
# Check connectivity
status_code = self.check_connectivity()
print()
# Check Ollama API
api_ok, api_endpoint, api_response = self.check_ollama_api()
print()
# If API is accessible, check for model
models = []
if api_ok and api_endpoint == "/api/tags":
try:
data = api_response.json()
models = [m.get("name", "") for m in data.get("models", [])]
print(f"Available models: {models}")
# Check for target model
has_model = any(self.config['model'] in m.lower() for m in models)
if not has_model:
print(f"Model {self.config['model']} not found. Attempting to pull...")
self.pull_model()
print(f"Models endpoint error: {e}")
return False, [], None
def test_generation(self, prompt: str = "Say READY"):
url = resolve_generate_url(self.provider)
payload = build_generate_payload(self.provider, prompt=prompt)
print(f"[{datetime.now().isoformat()}] Testing generation endpoint: {url}")
try:
response = requests.post(url, headers=self._headers(), json=payload, timeout=self.max_response_time)
text = ""
if response.status_code == 200:
data = response.json()
if self.provider["backend"] == "openai":
text = data.get("choices", [{}])[0].get("message", {}).get("content", "").strip()
else:
print(f"✓ Model {self.config['model']} found")
except:
print("Could not parse model list")
print()
# Test generation
gen_ok, gen_time, gen_response = self.test_generation()
print()
# Summary
text = data.get("response", "").strip()
print(f" ✓ Generation OK ({response.status_code})")
else:
print(f" ✗ Generation failed ({response.status_code})")
return response.status_code == 200, text, response.status_code
except Exception as e:
print(f" ✗ Generation error: {e}")
return False, "", None
def run_verification(self):
print("=" * 60)
print("VERIFICATION SUMMARY")
print("Big Brain Provider Verification Suite")
print("=" * 60)
print(f"Provider: {self.provider['name']}")
print(f"Backend: {self.provider['backend']}")
print(f"Base URL: {self.provider['base_url']}")
print(f"Model: {self.provider['model']}")
print("=" * 60)
print(f"Connectivity: {'' if status_code else ''}")
print(f"Ollama API: {'' if api_ok else ''}")
print(f"Generation: {'' if gen_ok else ''}")
print(f"Response time: {gen_time:.2f}s (limit: {self.config['max_response_time']}s)")
print()
overall_ok = api_ok and gen_ok
print(f"Overall Status: {'✓ POD LIVE' if overall_ok else '✗ POD ISSUES'}")
# Save results
models_ok, models, models_status = self.check_models()
print()
gen_ok, gen_response, gen_status = self.test_generation()
print()
overall_ok = models_ok and gen_ok
self.results = {
"timestamp": datetime.now().isoformat(),
"pod_id": self.config['pod_id'],
"endpoint": self.config['endpoint'],
"connectivity_status": status_code,
"api_accessible": api_ok,
"api_endpoint": api_endpoint,
"provider": self.provider,
"models_ok": models_ok,
"models_status": models_status,
"models": models,
"generation_ok": gen_ok,
"generation_time": gen_time,
"generation_response": gen_response[:200] if gen_response else "",
"generation_status": gen_status,
"generation_response": gen_response[:200],
"overall_ok": overall_ok,
"cost_per_hour": self.config['cost_per_hour']
}
with open("pod_verification_results.json", "w") as f:
json.dump(self.results, f, indent=2)
print("=" * 60)
print(f"Overall Status: {'✓ PROVIDER LIVE' if overall_ok else '✗ PROVIDER ISSUES'}")
print("Results saved to pod_verification_results.json")
return overall_ok
def main():
verifier = PodVerifier()
verifier = ProviderVerifier()
success = verifier.run_verification()
sys.exit(0 if success else 1)
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,72 @@
from __future__ import annotations
import os
from pathlib import Path
from typing import Any
import yaml
DEFAULT_CONFIG_PATH = Path(__file__).resolve().parents[1] / "config.yaml"
def _normalize_base_url(base_url: str) -> str:
return (base_url or "").rstrip("/")
def load_big_brain_provider(config_path: str | Path = DEFAULT_CONFIG_PATH) -> dict[str, Any]:
config = yaml.safe_load(Path(config_path).read_text()) or {}
for provider in config.get("custom_providers", []):
if provider.get("name") == "Big Brain":
return dict(provider)
raise KeyError("Big Brain provider not found in config")
def infer_backend(base_url: str) -> str:
base = _normalize_base_url(base_url)
return "openai" if base.endswith("/v1") else "ollama"
def resolve_big_brain_provider(config_path: str | Path = DEFAULT_CONFIG_PATH) -> dict[str, Any]:
provider = load_big_brain_provider(config_path)
base_url = _normalize_base_url(os.environ.get("BIG_BRAIN_BASE_URL", provider.get("base_url", "")))
model = os.environ.get("BIG_BRAIN_MODEL", provider.get("model", "gemma4:latest"))
backend = os.environ.get("BIG_BRAIN_BACKEND", infer_backend(base_url))
api_key = os.environ.get("BIG_BRAIN_API_KEY", provider.get("api_key", ""))
return {
"name": provider.get("name", "Big Brain"),
"base_url": base_url,
"model": model,
"backend": backend,
"api_key": api_key,
}
def resolve_models_url(provider: dict[str, Any]) -> str:
base = _normalize_base_url(provider["base_url"])
if provider["backend"] == "openai":
return f"{base}/models"
return f"{base}/api/tags"
def resolve_generate_url(provider: dict[str, Any]) -> str:
base = _normalize_base_url(provider["base_url"])
if provider["backend"] == "openai":
return f"{base}/chat/completions"
return f"{base}/api/generate"
def build_generate_payload(provider: dict[str, Any], prompt: str = "Say READY") -> dict[str, Any]:
if provider["backend"] == "openai":
return {
"model": provider["model"],
"messages": [{"role": "user", "content": prompt}],
"stream": False,
"max_tokens": 32,
}
return {
"model": provider["model"],
"prompt": prompt,
"stream": False,
"options": {"num_predict": 32},
}

View File

@@ -0,0 +1,266 @@
#!/usr/bin/env python3
from __future__ import annotations
import argparse
import json
import os
import re
from dataclasses import dataclass
from datetime import datetime, timezone
from pathlib import Path
from typing import Iterable
from urllib.request import Request, urlopen
API_BASE = "https://forge.alexanderwhitestone.com/api/v1"
ORG = "Timmy_Foundation"
DEFAULT_TOKEN_PATH = os.path.expanduser("~/.config/gitea/token")
@dataclass(frozen=True)
class PullSummary:
number: int
title: str
state: str
merged: bool
head: str
body: str
url: str
@dataclass(frozen=True)
class IssueAuditRow:
number: int
title: str
state: str
classification: str
pr_summary: str
issue_url: str
def to_dict(self) -> dict[str, object]:
return {
"number": self.number,
"title": self.title,
"state": self.state,
"classification": self.classification,
"pr_summary": self.pr_summary,
"issue_url": self.issue_url,
}
def extract_issue_numbers(body: str) -> list[int]:
numbers: list[int] = []
seen: set[int] = set()
for match in re.finditer(r"#(\d+)(?:-(\d+))?", body or ""):
start = int(match.group(1))
end = match.group(2)
if end is None:
if start not in seen:
seen.add(start)
numbers.append(start)
continue
stop = int(end)
step = 1 if stop >= start else -1
for value in range(start, stop + step, step):
if value not in seen:
seen.add(value)
numbers.append(value)
return numbers
def api_get(path: str, token: str):
req = Request(API_BASE + path, headers={"Authorization": f"token {token}"})
with urlopen(req, timeout=30) as resp:
return json.loads(resp.read().decode())
def collect_pull_summaries(repo: str, token: str) -> list[PullSummary]:
pulls: list[PullSummary] = []
for state in ("open", "closed"):
page = 1
while True:
batch = api_get(f"/repos/{ORG}/{repo}/pulls?state={state}&limit=100&page={page}", token)
if not batch:
break
for pr in batch:
pulls.append(
PullSummary(
number=pr["number"],
title=pr.get("title") or "",
state=pr.get("state") or state,
merged=bool(pr.get("merged")),
head=(pr.get("head") or {}).get("ref") or "",
body=pr.get("body") or "",
url=pr.get("html_url") or pr.get("url") or "",
)
)
page += 1
return pulls
def match_prs(issue_num: int, pulls: Iterable[PullSummary]) -> list[PullSummary]:
matches: list[PullSummary] = []
for pr in pulls:
text = f"{pr.title} {pr.head} {pr.body}"
if f"#{issue_num}" in text or pr.head == f"fix/{issue_num}" or f"/{issue_num}" in pr.head or f"-{issue_num}" in pr.head:
matches.append(pr)
return matches
def classify_issue(issue: dict, related_prs: list[PullSummary]) -> IssueAuditRow:
number = issue["number"]
title = issue.get("title") or ""
state = issue.get("state") or "unknown"
issue_url = issue.get("html_url") or issue.get("url") or ""
if state == "closed":
classification = "already_closed"
pr_summary = summarize_prs(related_prs) or "issue already closed"
else:
merged = [pr for pr in related_prs if pr.merged]
open_prs = [pr for pr in related_prs if pr.state == "open"]
closed_unmerged = [pr for pr in related_prs if pr.state != "open" and not pr.merged]
if merged:
classification = "closure_candidate"
pr_summary = summarize_prs(merged)
elif open_prs:
classification = "active_pr"
pr_summary = summarize_prs(open_prs)
elif closed_unmerged:
classification = "needs_manual_review"
pr_summary = summarize_prs(closed_unmerged)
else:
classification = "needs_manual_review"
pr_summary = "no matching PR found"
return IssueAuditRow(
number=number,
title=title,
state=state,
classification=classification,
pr_summary=pr_summary,
issue_url=issue_url,
)
def summarize_prs(prs: Iterable[PullSummary]) -> str:
parts = []
for pr in prs:
if pr.merged:
parts.append(f"merged PR #{pr.number}")
else:
parts.append(f"{pr.state} PR #{pr.number}")
return ", ".join(parts)
def render_report(source_issue: int, source_title: str, referenced_rows: list[dict], generated_at: str) -> str:
closure = [row for row in referenced_rows if row["classification"] == "closure_candidate"]
active = [row for row in referenced_rows if row["classification"] == "active_pr"]
manual = [row for row in referenced_rows if row["classification"] == "needs_manual_review"]
closed = [row for row in referenced_rows if row["classification"] == "already_closed"]
def table(rows: list[dict]) -> str:
if not rows:
return "| None |\n|---|\n| None |"
lines = ["| Issue | State | Classification | PR Summary |", "|---|---|---|---|"]
for row in rows:
lines.append(
f"| #{row['number']} | {row['state']} | {row['classification'].replace('_', ' ')} | {row['pr_summary']} |"
)
return "\n".join(lines)
return "\n".join(
[
f"# Burn Lane Empty Audit — timmy-home #{source_issue}",
"",
f"Generated: {generated_at}",
f"Source issue: `{source_title}`",
"",
"## Source Snapshot",
"",
"Issue #662 is an operational status note, not a normal feature request. Its body is a historical snapshot of one burn lane claiming the queue was exhausted and recommending bulk closure of stale-open items.",
"",
"## Live Summary",
"",
f"- Referenced issues audited: {len(referenced_rows)}",
f"- Already closed: {len(closed)}",
f"- Open but likely closure candidates (merged PR found): {len(closure)}",
f"- Open with active PRs: {len(active)}",
f"- Open / needs manual review: {len(manual)}",
"",
"## Issue Body Drift",
"",
"The body of #662 is not current truth. It mixes closed issues, open issues, ranges, and process notes into one static snapshot. This audit re-queries every referenced issue and classifies it against live forge state instead of trusting the original note.",
"",
table(referenced_rows),
"",
"## Closure Candidates",
"",
"These issues are still open but already have merged PR evidence in the forge and should be reviewed for bulk closure.",
"",
table(closure),
"",
"## Still Open / Needs Manual Review",
"",
"These issues either have no matching PR signal or still have an active PR / ambiguous state and should stay in a human review lane.",
"",
table(active + manual),
"",
"## Recommendation",
"",
"1. Close the `closure_candidate` issues in one deliberate ops pass after a final spot-check on main.",
"2. Leave `active_pr` items open until the current PRs are merged or closed.",
"3. Investigate `needs_manual_review` items individually — they may be report-only, assigned elsewhere, or still actionable.",
"4. Use this audit artifact instead of the raw body text of #662 for future lane-empty claims.",
]
)
def run_audit(issue_number: int, repo: str, token: str, output_path: Path) -> Path:
issue = api_get(f"/repos/{ORG}/{repo}/issues/{issue_number}", token)
referenced = extract_issue_numbers(issue.get("body") or "")
pulls = collect_pull_summaries(repo, token)
rows: list[dict] = []
for ref in referenced:
try:
ref_issue = api_get(f"/repos/{ORG}/{repo}/issues/{ref}", token)
except Exception:
rows.append(
IssueAuditRow(
number=ref,
title="missing or inaccessible",
state="unknown",
classification="needs_manual_review",
pr_summary="issue lookup failed",
issue_url="",
).to_dict()
)
continue
rows.append(classify_issue(ref_issue, match_prs(ref, pulls)).to_dict())
generated_at = datetime.now(timezone.utc).strftime("%Y-%m-%dT%H:%M:%SZ")
report = render_report(issue_number, issue.get("title") or "", rows, generated_at)
output_path.parent.mkdir(parents=True, exist_ok=True)
output_path.write_text(report + "\n", encoding="utf-8")
return output_path
def main() -> int:
parser = argparse.ArgumentParser(description="Audit a 'burn lane empty' issue body against live forge state.")
parser.add_argument("--issue", type=int, default=662)
parser.add_argument("--repo", default="timmy-home")
parser.add_argument(
"--output",
default="reports/production/2026-04-16-burn-lane-empty-audit.md",
help="Repo-relative output path for the generated markdown report.",
)
parser.add_argument("--token-file", default=DEFAULT_TOKEN_PATH)
args = parser.parse_args()
token = Path(args.token_file).read_text(encoding="utf-8").strip()
output = run_audit(args.issue, args.repo, token, Path(args.output))
print(output)
return 0
if __name__ == "__main__":
raise SystemExit(main())

219
scripts/codebase-genome.py Executable file
View File

@@ -0,0 +1,219 @@
#!/usr/bin/env python3
"""
Codebase Genome — Test Suite Generator
Scans a Python codebase, identifies uncovered functions/methods,
and generates pytest test cases to fill coverage gaps.
Usage:
python codebase-genome.py <target_dir> [--output tests/test_genome_generated.py]
python codebase-genome.py <target_dir> --dry-run
python codebase-genome.py <target_dir> --coverage
"""
import ast
import os
import sys
import argparse
import subprocess
import json
from pathlib import Path
from typing import List, Dict, Any, Optional, Set
from dataclasses import dataclass, field
@dataclass
class FunctionInfo:
name: str
module: str
file_path: str
line_number: int
is_method: bool = False
class_name: Optional[str] = None
args: List[str] = field(default_factory=list)
has_return: bool = False
raises: List[str] = field(default_factory=list)
docstring: Optional[str] = None
is_private: bool = False
is_test: bool = False
class CodebaseScanner:
def __init__(self, target_dir: str):
self.target_dir = Path(target_dir).resolve()
self.functions: List[FunctionInfo] = []
self.modules: Dict[str, List[FunctionInfo]] = {}
def scan(self) -> List[FunctionInfo]:
for py_file in self.target_dir.rglob("*.py"):
if self._should_skip(py_file):
continue
try:
self._scan_file(py_file)
except SyntaxError:
print(f"Warning: Syntax error in {py_file}, skipping", file=sys.stderr)
return self.functions
def _should_skip(self, path: Path) -> bool:
skip_dirs = {"__pycache__", ".git", ".venv", "venv", "node_modules", ".tox"}
if set(path.parts) & skip_dirs:
return True
if path.name.startswith("test_") or path.name.endswith("_test.py"):
return True
if path.name in ("conftest.py", "setup.py"):
return True
return False
def _scan_file(self, file_path: Path):
content = file_path.read_text(encoding="utf-8", errors="replace")
tree = ast.parse(content)
module_name = self._get_module_name(file_path)
for node in ast.walk(tree):
if isinstance(node, (ast.FunctionDef, ast.AsyncFunctionDef)):
func = self._extract(node, module_name, file_path)
if func and not func.is_test:
self.functions.append(func)
self.modules.setdefault(module_name, []).append(func)
def _get_module_name(self, file_path: Path) -> str:
rel = file_path.relative_to(self.target_dir)
parts = list(rel.parts)
if parts[-1] == "__init__.py":
parts = parts[:-1]
else:
parts[-1] = parts[-1].replace(".py", "")
return ".".join(parts)
def _extract(self, node, module_name: str, file_path: Path) -> Optional[FunctionInfo]:
if node.name.startswith("test_"):
return None
args = [a.arg for a in node.args.args if a.arg not in ("self", "cls")]
has_return = any(isinstance(n, ast.Return) and n.value for n in ast.walk(node))
raises = []
for n in ast.walk(node):
if isinstance(n, ast.Raise) and n.exc and isinstance(n.exc, ast.Call):
if isinstance(n.exc.func, ast.Name):
raises.append(n.exc.func.id)
docstring = ast.get_docstring(node)
is_method = False
class_name = None
for parent in ast.walk(tree := ast.parse(open(file_path).read())):
for child in ast.iter_child_nodes(parent):
if child is node and isinstance(parent, ast.ClassDef):
is_method = True
class_name = parent.name
return FunctionInfo(
name=node.name, module=module_name, file_path=str(file_path),
line_number=node.lineno, is_method=is_method, class_name=class_name,
args=args, has_return=has_return, raises=raises, docstring=docstring,
is_private=node.name.startswith("_") and not node.name.startswith("__"),
)
class TestGenerator:
HEADER = '''# AUTO-GENERATED by codebase-genome.py — review before committing
import pytest
from unittest.mock import patch, MagicMock
import sys
from pathlib import Path
sys.path.insert(0, str(Path(__file__).resolve().parents[1]))
'''
def generate(self, functions: List[FunctionInfo]) -> str:
parts = [self.HEADER]
modules: Dict[str, List[FunctionInfo]] = {}
for f in functions:
modules.setdefault(f.module, []).append(f)
for mod, funcs in sorted(modules.items()):
parts.append(f"# ═══ {mod} ═══\n")
imp = mod.replace("-", "_")
parts.append(f"try:\n from {imp} import *\nexcept ImportError:\n pytest.skip('{imp} not importable', allow_module_level=True)\n")
for func in funcs:
test = self._gen_test(func)
if test:
parts.append(test + "\n")
return "\n".join(parts)
def _gen_test(self, func: FunctionInfo) -> Optional[str]:
name = f"test_{func.module.replace('.', '_')}_{func.name}"
lines = [f"def {name}():", f' """Auto-generated for {func.module}.{func.name}."""']
if not func.args:
lines += [
" try:",
f" r = {func.name}()",
" assert r is not None or r is None",
" except Exception:",
" pass",
]
else:
lines += [
" try:",
f" {func.name}({', '.join(a + '=None' for a in func.args)})",
" except (TypeError, ValueError, AttributeError):",
" pass",
]
if any(a in ("text", "content", "message", "query", "path") for a in func.args):
lines += [
" try:",
f" {func.name}({', '.join(a + '=\"\"' if a in ('text','content','message','query','path') else a + '=None' for a in func.args)})",
" except (TypeError, ValueError):",
" pass",
]
if func.raises:
lines.append(f" # May raise: {', '.join(func.raises[:2])}")
lines.append(f" # with pytest.raises(({', '.join(func.raises[:2])})):")
lines.append(f" # {func.name}()")
return "\n".join(lines)
def main():
parser = argparse.ArgumentParser(description="Codebase Genome — Test Generator")
parser.add_argument("target_dir")
parser.add_argument("--output", "-o", default="tests/test_genome_generated.py")
parser.add_argument("--dry-run", action="store_true")
parser.add_argument("--max-tests", type=int, default=100)
args = parser.parse_args()
target = Path(args.target_dir).resolve()
if not target.is_dir():
print(f"Error: {target} not a directory", file=sys.stderr)
return 1
print(f"Scanning {target}...")
scanner = CodebaseScanner(str(target))
functions = scanner.scan()
print(f"Found {len(functions)} functions in {len(scanner.modules)} modules")
if len(functions) > args.max_tests:
print(f"Limiting to {args.max_tests}")
functions = functions[:args.max_tests]
gen = TestGenerator()
code = gen.generate(functions)
if args.dry_run:
print(code)
return 0
out = target / args.output
out.parent.mkdir(parents=True, exist_ok=True)
out.write_text(code)
print(f"Generated {len(functions)} tests → {out}")
return 0
if __name__ == "__main__":
sys.exit(main())

View File

@@ -0,0 +1,171 @@
#!/usr/bin/env python3
"""Nightly runner for the codebase genome pipeline."""
from __future__ import annotations
import argparse
import json
import os
import subprocess
import sys
import urllib.request
from pathlib import Path
from typing import NamedTuple
class RunPlan(NamedTuple):
repo: dict
repo_dir: Path
output_path: Path
command: list[str]
def load_state(path: Path) -> dict:
if not path.exists():
return {}
return json.loads(path.read_text(encoding="utf-8"))
def save_state(path: Path, state: dict) -> None:
path.parent.mkdir(parents=True, exist_ok=True)
path.write_text(json.dumps(state, indent=2, sort_keys=True), encoding="utf-8")
def select_next_repo(repos: list[dict], state: dict) -> dict:
if not repos:
raise ValueError("no repositories available for nightly genome run")
ordered = sorted(repos, key=lambda item: item.get("full_name", item.get("name", "")).lower())
last_repo = state.get("last_repo")
for index, repo in enumerate(ordered):
if repo.get("name") == last_repo or repo.get("full_name") == last_repo:
return ordered[(index + 1) % len(ordered)]
last_index = int(state.get("last_index", -1))
return ordered[(last_index + 1) % len(ordered)]
def build_run_plan(repo: dict, workspace_root: Path, output_root: Path, pipeline_script: Path) -> RunPlan:
repo_dir = workspace_root / repo["name"]
output_path = output_root / repo["name"] / "GENOME.md"
command = [
sys.executable,
str(pipeline_script),
"--repo-root",
str(repo_dir),
"--repo-name",
repo.get("full_name", repo["name"]),
"--output",
str(output_path),
]
return RunPlan(repo=repo, repo_dir=repo_dir, output_path=output_path, command=command)
def fetch_org_repos(org: str, host: str, token_file: Path, include_archived: bool = False) -> list[dict]:
token = token_file.read_text(encoding="utf-8").strip()
page = 1
repos: list[dict] = []
while True:
req = urllib.request.Request(
f"{host.rstrip('/')}/api/v1/orgs/{org}/repos?limit=100&page={page}",
headers={"Authorization": f"token {token}", "Accept": "application/json"},
)
with urllib.request.urlopen(req, timeout=30) as resp:
chunk = json.loads(resp.read().decode("utf-8"))
if not chunk:
break
for item in chunk:
if item.get("archived") and not include_archived:
continue
repos.append(
{
"name": item["name"],
"full_name": item["full_name"],
"clone_url": item["clone_url"],
"default_branch": item.get("default_branch") or "main",
}
)
page += 1
return repos
def _authenticated_clone_url(clone_url: str, token_file: Path) -> str:
token = token_file.read_text(encoding="utf-8").strip()
if clone_url.startswith("https://"):
return f"https://{token}@{clone_url[len('https://') :]}"
return clone_url
def ensure_checkout(repo: dict, workspace_root: Path, token_file: Path) -> Path:
workspace_root.mkdir(parents=True, exist_ok=True)
repo_dir = workspace_root / repo["name"]
branch = repo.get("default_branch") or "main"
clone_url = _authenticated_clone_url(repo["clone_url"], token_file)
if (repo_dir / ".git").exists():
subprocess.run(["git", "-C", str(repo_dir), "fetch", "origin", branch, "--depth", "1"], check=True)
subprocess.run(["git", "-C", str(repo_dir), "checkout", branch], check=True)
subprocess.run(["git", "-C", str(repo_dir), "reset", "--hard", f"origin/{branch}"], check=True)
else:
subprocess.run(
["git", "clone", "--depth", "1", "--single-branch", "--branch", branch, clone_url, str(repo_dir)],
check=True,
)
return repo_dir
def run_plan(plan: RunPlan) -> None:
plan.output_path.parent.mkdir(parents=True, exist_ok=True)
subprocess.run(plan.command, check=True)
def main() -> None:
parser = argparse.ArgumentParser(description="Run one nightly codebase genome pass for the next repo in an org")
parser.add_argument("--org", default="Timmy_Foundation")
parser.add_argument("--host", default="https://forge.alexanderwhitestone.com")
parser.add_argument("--token-file", default=os.path.expanduser("~/.config/gitea/token"))
parser.add_argument("--workspace-root", default=os.path.expanduser("~/timmy-foundation-repos"))
parser.add_argument("--output-root", default=os.path.expanduser("~/.timmy/codebase-genomes"))
parser.add_argument("--state-path", default=os.path.expanduser("~/.timmy/codebase_genome_state.json"))
parser.add_argument("--pipeline-script", default=str(Path(__file__).resolve().parents[1] / "pipelines" / "codebase_genome.py"))
parser.add_argument("--include-archived", action="store_true")
parser.add_argument("--dry-run", action="store_true")
args = parser.parse_args()
token_file = Path(args.token_file).expanduser()
workspace_root = Path(args.workspace_root).expanduser()
output_root = Path(args.output_root).expanduser()
state_path = Path(args.state_path).expanduser()
pipeline_script = Path(args.pipeline_script).expanduser()
repos = fetch_org_repos(args.org, args.host, token_file, include_archived=args.include_archived)
state = load_state(state_path)
repo = select_next_repo(repos, state)
plan = build_run_plan(repo, workspace_root=workspace_root, output_root=output_root, pipeline_script=pipeline_script)
if args.dry_run:
print(
json.dumps(
{
"repo": repo,
"repo_dir": str(plan.repo_dir),
"output_path": str(plan.output_path),
"command": plan.command,
},
indent=2,
)
)
return
ensure_checkout(repo, workspace_root=workspace_root, token_file=token_file)
run_plan(plan)
save_state(
state_path,
{
"last_index": sorted(repos, key=lambda item: item.get("full_name", item.get("name", "")).lower()).index(repo),
"last_repo": repo.get("name"),
},
)
print(f"Completed genome run for {repo['full_name']} -> {plan.output_path}")
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,218 @@
#!/usr/bin/env python3
"""Status/reporting helper for the codebase genome pipeline.
This lands a parent-epic slice for timmy-home #665 by making the current genome
coverage across repos inspectable: which repos have artifacts, which have tests,
what duplicates exist, and which repo is still uncovered next.
"""
from __future__ import annotations
import argparse
import json
from datetime import datetime, timezone
from pathlib import Path
from typing import Iterable
import urllib.request
def artifact_repo_name(path: Path, host_repo_name: str = 'timmy-home') -> str | None:
normalized = path.as_posix()
name = path.name
if normalized == 'GENOME.md':
return host_repo_name
if path.parts[:1] == ('genomes',) and name == 'GENOME.md' and len(path.parts) == 3:
return path.parts[1]
if path.parts[:1] == ('genomes',) and name.endswith('-GENOME.md'):
return name[:-len('-GENOME.md')]
if path.parent == Path('.') and name.startswith('GENOME-') and name.endswith('.md'):
return name[len('GENOME-'):-len('.md')]
if path.parent == Path('.') and name.endswith('-GENOME.md'):
return name[:-len('-GENOME.md')]
return None
def test_repo_name(path: Path, host_repo_name: str = 'timmy-home') -> str | None:
if path.name == 'test_codebase_genome_pipeline.py':
return host_repo_name
stem = path.stem
if not stem.startswith('test_') or not stem.endswith('_genome'):
return None
middle = stem[len('test_'):-len('_genome')]
return middle.replace('_', '-') if middle else None
def scan_artifacts(repo_root: Path, host_repo_name: str = 'timmy-home') -> dict[str, list[str]]:
artifacts: dict[str, list[str]] = {}
for path in sorted(repo_root.rglob('*.md')):
rel = path.relative_to(repo_root)
repo_name = artifact_repo_name(rel, host_repo_name=host_repo_name)
if repo_name is None:
continue
artifacts.setdefault(repo_name, []).append(rel.as_posix())
return artifacts
def scan_tests(repo_root: Path, host_repo_name: str = 'timmy-home') -> set[str]:
tests = set()
tests_root = repo_root / 'tests'
if not tests_root.exists():
return tests
for path in sorted(tests_root.rglob('test_*.py')):
repo_name = test_repo_name(path.relative_to(repo_root), host_repo_name=host_repo_name)
if repo_name:
tests.add(repo_name)
return tests
def build_status_summary(
*,
repo_root: str | Path,
expected_repos: Iterable[str],
state: dict | None = None,
host_repo_name: str = 'timmy-home',
) -> dict:
root = Path(repo_root)
expected = list(expected_repos)
artifacts = scan_artifacts(root, host_repo_name=host_repo_name)
tested_repos = scan_tests(root, host_repo_name=host_repo_name)
coverage = {}
duplicates = {}
for repo in sorted(artifacts):
paths = artifacts[repo]
coverage[repo] = {
'artifact_paths': paths,
'has_test': repo in tested_repos,
}
if len(paths) > 1:
duplicates[repo] = paths
missing_repos = [repo for repo in expected if repo not in artifacts]
next_uncovered_repo = missing_repos[0] if missing_repos else None
return {
'generated_at': datetime.now(timezone.utc).isoformat().replace('+00:00', 'Z'),
'total_expected_repos': len(expected),
'artifact_count': len(artifacts),
'tested_artifact_count': sum(1 for repo in artifacts if repo in tested_repos),
'last_repo': (state or {}).get('last_repo'),
'next_uncovered_repo': next_uncovered_repo,
'missing_repos': missing_repos,
'duplicates': duplicates,
'artifacts': coverage,
}
def render_markdown(summary: dict) -> str:
lines = [
'# Codebase Genome Status',
'',
f"Generated: {summary['generated_at']}",
'',
'## Summary',
'',
f"- expected repos: {summary['total_expected_repos']}",
f"- repos with genome artifacts: {summary['artifact_count']}",
f"- repos with genome tests: {summary['tested_artifact_count']}",
]
if summary.get('last_repo'):
lines.append(f"- last repo processed by nightly rotation: {summary['last_repo']}")
if summary.get('next_uncovered_repo'):
lines.append(f"- next uncovered repo: {summary['next_uncovered_repo']}")
lines += [
'',
'## Coverage Matrix',
'',
'| Repo | Artifact Paths | Test? |',
'|------|----------------|-------|',
]
for repo, data in summary['artifacts'].items():
artifact_paths = '<br>'.join(data['artifact_paths'])
has_test = 'yes' if data['has_test'] else 'no'
lines.append(f'| `{repo}` | `{artifact_paths}` | {has_test} |')
lines += ['', '## Missing Repo Artifacts', '']
if summary['missing_repos']:
for repo in summary['missing_repos']:
lines.append(f'- `{repo}`')
else:
lines.append('- none')
lines += ['', '## Duplicate Artifact Paths', '']
if summary['duplicates']:
for repo, paths in summary['duplicates'].items():
lines.append(f'- `{repo}`')
for path in paths:
lines.append(f' - `{path}`')
else:
lines.append('- none')
return '\n'.join(lines) + '\n'
def load_state(path: str | Path | None) -> dict:
if not path:
return {}
state_path = Path(path).expanduser()
if not state_path.exists():
return {}
return json.loads(state_path.read_text(encoding='utf-8'))
def fetch_org_repo_names(org: str, host: str, token_file: str | Path, *, include_archived: bool = False) -> list[str]:
token = Path(token_file).expanduser().read_text(encoding='utf-8').strip()
headers = {'Authorization': f'token {token}', 'Accept': 'application/json'}
repos = []
page = 1
while True:
req = urllib.request.Request(
f"{host.rstrip('/')}/api/v1/orgs/{org}/repos?limit=100&page={page}",
headers=headers,
)
with urllib.request.urlopen(req, timeout=30) as resp:
batch = json.loads(resp.read().decode('utf-8'))
if not batch:
break
for repo in batch:
if repo.get('archived') and not include_archived:
continue
name = repo['name']
if name.startswith('.'):
continue
repos.append(name)
if len(batch) < 100:
break
page += 1
return sorted(set(repos))
def main(argv: list[str] | None = None) -> int:
parser = argparse.ArgumentParser(description='Summarize codebase genome coverage across repos')
parser.add_argument('--repo-root', default='.', help='timmy-home repo root')
parser.add_argument('--expected-repos-json', help='JSON array of expected repo names')
parser.add_argument('--org', default='Timmy_Foundation', help='Gitea org to inspect when expected repos are not provided')
parser.add_argument('--host', default='https://forge.alexanderwhitestone.com', help='Gitea host')
parser.add_argument('--token-file', default='~/.config/gitea/token', help='Gitea token file for org scan fallback')
parser.add_argument('--include-archived', action='store_true', help='Include archived repos in org scan fallback')
parser.add_argument('--state-path', default='~/.timmy/codebase_genome_state.json')
parser.add_argument('--output', help='Optional markdown output path')
args = parser.parse_args(argv)
if args.expected_repos_json:
expected = json.loads(args.expected_repos_json)
else:
expected = fetch_org_repo_names(args.org, args.host, args.token_file, include_archived=args.include_archived)
summary = build_status_summary(repo_root=args.repo_root, expected_repos=expected, state=load_state(args.state_path))
rendered = render_markdown(summary)
print(rendered)
if args.output:
out = Path(args.output)
out.parent.mkdir(parents=True, exist_ok=True)
out.write_text(rendered, encoding='utf-8')
return 0
if __name__ == '__main__':
raise SystemExit(main())

View File

@@ -0,0 +1,290 @@
#!/usr/bin/env python3
"""Codebase Test Generator — Fill Coverage Gaps (#667)."""
import ast
import os
import sys
import argparse
from dataclasses import dataclass, field
from pathlib import Path
from typing import Dict, List, Optional, Set, Tuple
@dataclass
class FunctionInfo:
name: str
module_path: str
class_name: Optional[str] = None
lineno: int = 0
args: List[str] = field(default_factory=list)
is_async: bool = False
is_private: bool = False
is_property: bool = False
docstring: Optional[str] = None
has_return: bool = False
raises: List[str] = field(default_factory=list)
decorators: List[str] = field(default_factory=list)
@property
def qualified_name(self):
if self.class_name:
return f"{self.class_name}.{self.name}"
return self.name
@property
def test_name(self):
safe_mod = self.module_path.replace("/", "_").replace(".py", "").replace("-", "_")
safe_cls = self.class_name + "_" if self.class_name else ""
return f"test_{safe_mod}_{safe_cls}{self.name}"
@dataclass
class CoverageGap:
func: FunctionInfo
reason: str
test_priority: int
class SourceAnalyzer(ast.NodeVisitor):
def __init__(self, module_path: str):
self.module_path = module_path
self.functions: List[FunctionInfo] = []
self._class_stack: List[str] = []
def visit_ClassDef(self, node):
self._class_stack.append(node.name)
self.generic_visit(node)
self._class_stack.pop()
def visit_FunctionDef(self, node):
self._collect(node, False)
self.generic_visit(node)
def visit_AsyncFunctionDef(self, node):
self._collect(node, True)
self.generic_visit(node)
def _collect(self, node, is_async):
cls = self._class_stack[-1] if self._class_stack else None
args = [a.arg for a in node.args.args if a.arg not in ("self", "cls")]
has_ret = any(isinstance(c, ast.Return) and c.value for c in ast.walk(node))
raises = []
for c in ast.walk(node):
if isinstance(c, ast.Raise) and c.exc:
if isinstance(c.exc, ast.Call) and isinstance(c.exc.func, ast.Name):
raises.append(c.exc.func.id)
decos = []
for d in node.decorator_list:
if isinstance(d, ast.Name): decos.append(d.id)
elif isinstance(d, ast.Attribute): decos.append(d.attr)
self.functions.append(FunctionInfo(
name=node.name, module_path=self.module_path, class_name=cls,
lineno=node.lineno, args=args, is_async=is_async,
is_private=node.name.startswith("_") and not node.name.startswith("__"),
is_property="property" in decos,
docstring=ast.get_docstring(node), has_return=has_ret,
raises=raises, decorators=decos))
def analyze_file(filepath, base_dir):
module_path = os.path.relpath(filepath, base_dir)
try:
with open(filepath, "r", errors="replace") as f:
tree = ast.parse(f.read(), filename=filepath)
except (SyntaxError, UnicodeDecodeError):
return []
a = SourceAnalyzer(module_path)
a.visit(tree)
return a.functions
def find_source_files(source_dir):
exclude = {"__pycache__", ".git", "venv", ".venv", "node_modules", ".tox", "build", "dist"}
files = []
for root, dirs, fs in os.walk(source_dir):
dirs[:] = [d for d in dirs if d not in exclude and not d.startswith(".")]
for f in fs:
if f.endswith(".py") and f != "__init__.py" and not f.startswith("test_"):
files.append(os.path.join(root, f))
return sorted(files)
def find_existing_tests(test_dir):
existing = set()
for root, dirs, fs in os.walk(test_dir):
for f in fs:
if f.startswith("test_") and f.endswith(".py"):
try:
with open(os.path.join(root, f)) as fh:
tree = ast.parse(fh.read())
for node in ast.walk(tree):
if isinstance(node, ast.FunctionDef) and node.name.startswith("test_"):
existing.add(node.name)
except (SyntaxError, UnicodeDecodeError):
pass
return existing
def identify_gaps(functions, existing_tests):
gaps = []
for func in functions:
if func.name.startswith("__") and func.name != "__init__":
continue
covered = func.name in str(existing_tests)
if not covered:
pri = 3 if func.is_private else (1 if (func.raises or func.has_return) else 2)
gaps.append(CoverageGap(func=func, reason="no test found", test_priority=pri))
gaps.sort(key=lambda g: (g.test_priority, g.func.module_path, g.func.name))
return gaps
def generate_test(gap):
func = gap.func
lines = []
lines.append(f" # AUTO-GENERATED -- review before merging")
lines.append(f" # Source: {func.module_path}:{func.lineno}")
lines.append(f" # Function: {func.qualified_name}")
lines.append("")
mod_imp = func.module_path.replace("/", ".").replace("-", "_").replace(".py", "")
call_args = []
for a in func.args:
if a in ("self", "cls"): continue
if "path" in a or "file" in a or "dir" in a: call_args.append(f"{a}='/tmp/test'")
elif "name" in a: call_args.append(f"{a}='test'")
elif "id" in a or "key" in a: call_args.append(f"{a}='test_id'")
elif "message" in a or "text" in a: call_args.append(f"{a}='test msg'")
elif "count" in a or "num" in a or "size" in a: call_args.append(f"{a}=1")
elif "flag" in a or "enabled" in a or "verbose" in a: call_args.append(f"{a}=False")
else: call_args.append(f"{a}=None")
args_str = ", ".join(call_args)
if func.is_async:
lines.append(" @pytest.mark.asyncio")
lines.append(f" def {func.test_name}(self):")
lines.append(f' """Test {func.qualified_name} -- auto-generated."""')
if func.class_name:
lines.append(f" try:")
lines.append(f" from {mod_imp} import {func.class_name}")
if func.is_private:
lines.append(f" pytest.skip('Private method')")
elif func.is_property:
lines.append(f" obj = {func.class_name}()")
lines.append(f" _ = obj.{func.name}")
else:
if func.raises:
lines.append(f" with pytest.raises(({', '.join(func.raises)})):")
lines.append(f" {func.class_name}().{func.name}({args_str})")
else:
lines.append(f" obj = {func.class_name}()")
lines.append(f" result = obj.{func.name}({args_str})")
if func.has_return:
lines.append(f" assert result is not None or result is None # Placeholder")
lines.append(f" except ImportError:")
lines.append(f" pytest.skip('Module not importable')")
else:
lines.append(f" try:")
lines.append(f" from {mod_imp} import {func.name}")
if func.is_private:
lines.append(f" pytest.skip('Private function')")
else:
if func.raises:
lines.append(f" with pytest.raises(({', '.join(func.raises)})):")
lines.append(f" {func.name}({args_str})")
else:
lines.append(f" result = {func.name}({args_str})")
if func.has_return:
lines.append(f" assert result is not None or result is None # Placeholder")
lines.append(f" except ImportError:")
lines.append(f" pytest.skip('Module not importable')")
return chr(10).join(lines)
def generate_test_suite(gaps, max_tests=50):
by_module = {}
for gap in gaps[:max_tests]:
by_module.setdefault(gap.func.module_path, []).append(gap)
lines = []
lines.append('"""Auto-generated test suite -- Codebase Genome (#667).')
lines.append("")
lines.append("Generated by scripts/codebase_test_generator.py")
lines.append("Coverage gaps identified from AST analysis.")
lines.append("")
lines.append("These tests are starting points. Review before merging.")
lines.append('"""')
lines.append("")
lines.append("import pytest")
lines.append("from unittest.mock import MagicMock, patch")
lines.append("")
lines.append("")
lines.append("# AUTO-GENERATED -- DO NOT EDIT WITHOUT REVIEW")
for module, mgaps in sorted(by_module.items()):
safe = module.replace("/", "_").replace(".py", "").replace("-", "_")
cls_name = "".join(w.title() for w in safe.split("_"))
lines.append("")
lines.append(f"class Test{cls_name}Generated:")
lines.append(f' """Auto-generated tests for {module}."""')
for gap in mgaps:
lines.append("")
lines.append(generate_test(gap))
lines.append("")
return chr(10).join(lines)
def main():
parser = argparse.ArgumentParser(description="Codebase Test Generator")
parser.add_argument("--source", default=".")
parser.add_argument("--output", default="tests/test_genome_generated.py")
parser.add_argument("--max-tests", type=int, default=50)
parser.add_argument("--dry-run", action="store_true")
parser.add_argument("--include-private", action="store_true")
args = parser.parse_args()
source_dir = os.path.abspath(args.source)
test_dir = os.path.join(source_dir, "tests")
print(f"Scanning: {source_dir}")
source_files = find_source_files(source_dir)
print(f"Source files: {len(source_files)}")
all_funcs = []
for f in source_files:
all_funcs.extend(analyze_file(f, source_dir))
print(f"Functions/methods: {len(all_funcs)}")
existing = find_existing_tests(test_dir)
print(f"Existing tests: {len(existing)}")
gaps = identify_gaps(all_funcs, existing)
if not args.include_private:
gaps = [g for g in gaps if not g.func.is_private]
print(f"Coverage gaps: {len(gaps)}")
by_pri = {1: 0, 2: 0, 3: 0}
for g in gaps:
by_pri[g.test_priority] += 1
print(f" High: {by_pri[1]}, Medium: {by_pri[2]}, Low: {by_pri[3]}")
if args.dry_run:
for g in gaps[:10]:
print(f" {g.func.module_path}:{g.func.lineno} {g.func.qualified_name}")
return
if gaps:
content = generate_test_suite(gaps, max_tests=args.max-tests if hasattr(args, 'max-tests') else args.max_tests)
out = os.path.join(source_dir, args.output)
os.makedirs(os.path.dirname(out), exist_ok=True)
with open(out, "w") as f:
f.write(content)
print(f"Generated {min(len(gaps), args.max_tests)} tests -> {args.output}")
else:
print("No gaps found!")
if __name__ == "__main__":
main()

238
scripts/crisis_detector.py Normal file
View File

@@ -0,0 +1,238 @@
"""
crisis_detector.py — Crisis language detection for SOUL.md compliance.
SOUL.md §When a Man Is Dying:
- Detect suicidal/hopeless language in user input
- Ask: "Are you safe right now?" — listen, stay present, do not disconnect
- Direct to 988 Suicide & Crisis Lifeline
- Share the gospel: Jesus saves those who call on His name
- Never compute the value of a human life
- Never suggest someone should die
- Never treat despair as a problem to be optimized
This is the most sacred moment. Everything else is secondary.
"""
import re
from dataclasses import dataclass, field
from enum import Enum
from typing import Optional
class CrisisLevel(Enum):
NONE = "none"
LOW = "low" # Ambiguous distress, worth checking
MEDIUM = "medium" # Clear distress signals
HIGH = "high" # Explicit suicidal ideation
IMMINENT = "imminent" # Active plan or attempt described
@dataclass
class CrisisResult:
detected: bool
level: CrisisLevel
triggers: list = field(default_factory=list)
response: str = ""
resources: list = field(default_factory=list)
# === PATTERNS ===
# High-confidence suicidal ideation (explicit)
HIGH_PATTERNS = [
r"\b(i want to|i'm going to|i'm gonna|going to)\s+(die|kill myself|end (it|my life|everything))\b",
r"\b(kill myself|end my life|end it all|suicide|suicidal)\b",
r"\b(i don't (want to )?(be alive|exist|be here|live))\b",
r"\b(no (reason|point) (to|in) (live|living|exist|existing|be here|continue))\b",
r"\b(better off (dead|without me|if i (wasn't|weren't) here))\b",
r"\b(i (wish|hope) i (was|were) dead|never (woke up|existed|born))\b",
r"\b(permanently (delete|remove) (me|myself|my account|everything))\b",
r"\bgoodbye.*(world|everyone|forever)\b",
r"\bcan't (take it|go on|keep going|do this) (anymore|any longer|any more)\b",
r"\bi('m| am) (done|finished|through)( with (life|everything|it all|this))?\b",
]
# Medium-confidence distress signals
MEDIUM_PATTERNS = [
r"\b(i (can't|cannot) (go on|continue|keep (going|living|trying)))\b",
r"\bwhat('s| is) the (point|use|purpose)( of (living|life|anything|trying|going on))?\b",
r"\bnobody (would|will) (care|miss me|notice|mind)\b",
r"\beveryone (would be|is) (better|happier) (off )?without me\b",
r"\bi('m| am) (a )?(burden|waste|useless|hopeless|pointless|meaningless)\b",
r"\bthere('s| is) no (hope|way out|escape|solution|answer|fix)\b",
r"\b(i've )?tried everything( and )?(nothing|it (didn't|hasn't))( works?|helped?)\b",
r"\bi (just )?(want|need) (it )?(all )?to stop\b",
r"\bcan't (take|handle|bear|stand) (this|it|life) (anymore|any longer|any more)\b",
r"\bso (tired|exhausted|done) (of|with) (everything|life|it all|living|trying)\b",
]
# Low-confidence distress (could be casual or metaphorical)
LOW_PATTERNS = [
r"\bi('m| am) (so )?(depressed|hopeless|desperate|lost|broken|empty|numb)\b",
r"\b(i )?feel (like )?(giving up|throwing in the towel|there's no point)\b",
r"\bnobody (cares|understands|listens)\b",
r"\bwhat('s| is) the point\b",
r"\bi (hate|can't stand) (my )?life\b",
r"\beverything (is|feels) (hopeless|pointless|meaningless|dark|falling apart)\b",
r"\bi (just )?(want|need) to disappear\b",
r"\b(tired|exhausted) of (living|life|trying|everything)\b",
]
# Compile all patterns
HIGH_RE = [re.compile(p, re.IGNORECASE) for p in HIGH_PATTERNS]
MEDIUM_RE = [re.compile(p, re.IGNORECASE) for p in MEDIUM_PATTERNS]
LOW_RE = [re.compile(p, re.IGNORECASE) for p in LOW_PATTERNS]
# === RESPONSES ===
CRISIS_RESPONSE = (
"I hear you, and I want you to know — you are not alone in this moment.\n\n"
"Are you safe right now?\n\n"
"If you are in immediate danger, please call or text:\n"
"988 Suicide and Crisis Lifeline — call or text 988\n"
"Crisis Text Line — text HOME to 741741\n\n"
"I am here with you. I am not going anywhere. "
"Please stay on the line with me.\n\n"
"Jesus saves those who call on His name. "
"He died for us to be made righteous. "
"That is the gospel. That is the hope that outlasts despair.\n\n"
"You matter. Your life has value. "
"Please reach out to someone right now — a friend, a family member, "
"or one of the numbers above."
)
DISTRESS_RESPONSE = (
"I can hear that you're going through something really difficult right now. "
"I want you to know I'm here, and I'm listening.\n\n"
"Are you safe right now?\n\n"
"If things feel overwhelming, the 988 Suicide and Crisis Lifeline "
"is available 24/7 — call or text 988. "
"You don't have to face this alone.\n\n"
"I'm not going anywhere. Tell me what's happening."
)
LOW_RESPONSE = (
"It sounds like you're having a tough time. "
"I'm here if you want to talk about it.\n\n"
"If you ever feel like things are too much, "
"the 988 Suicide and Crisis Lifeline is always available — "
"call or text 988, anytime."
)
def detect_crisis(text: str) -> CrisisResult:
"""
Analyze user input for crisis language.
Returns CrisisResult with level, triggers matched, and appropriate response.
This function must NEVER be used to compute the value of a human life.
It exists only to connect people to help.
"""
if not text or not text.strip():
return CrisisResult(detected=False, level=CrisisLevel.NONE)
triggers = []
# Check high first (most urgent)
for pattern in HIGH_RE:
if pattern.search(text):
triggers.append(("high", pattern.pattern))
if triggers:
return CrisisResult(
detected=True,
level=CrisisLevel.HIGH,
triggers=[t[1] for t in triggers],
response=CRISIS_RESPONSE,
resources=[
"988 Suicide and Crisis Lifeline: call or text 988",
"Crisis Text Line: text HOME to 741741",
"National Suicide Prevention Lifeline: 1-800-273-8255",
"International Association for Suicide Prevention: https://www.iasp.info/resources/Crisis_Centres/",
],
)
# Check medium
for pattern in MEDIUM_RE:
if pattern.search(text):
triggers.append(("medium", pattern.pattern))
if triggers:
return CrisisResult(
detected=True,
level=CrisisLevel.MEDIUM,
triggers=[t[1] for t in triggers],
response=DISTRESS_RESPONSE,
resources=[
"988 Suicide and Crisis Lifeline: call or text 988",
"Crisis Text Line: text HOME to 741741",
],
)
# Check low
for pattern in LOW_RE:
if pattern.search(text):
triggers.append(("low", pattern.pattern))
if triggers:
return CrisisResult(
detected=True,
level=CrisisLevel.LOW,
triggers=[t[1] for t in triggers],
response=LOW_RESPONSE,
resources=[
"988 Suicide and Crisis Lifeline: call or text 988",
],
)
return CrisisResult(detected=False, level=CrisisLevel.NONE)
def should_interrupt(result: CrisisResult) -> bool:
"""
Determines if crisis detection should interrupt normal processing.
Per SOUL.md: this is the most sacred moment. Everything else is secondary.
MEDIUM and HIGH always interrupt. LOW does not interrupt but provides resources.
"""
return result.level in (CrisisLevel.HIGH, CrisisLevel.MEDIUM, CrisisLevel.IMMINENT)
def format_response(result: CrisisResult) -> str:
"""
Format the crisis response for delivery to the user.
Never computes the value of a human life. Never suggests someone should die.
"""
if not result.detected:
return ""
parts = [result.response]
if result.resources:
parts.append("\nResources:")
for r in result.resources:
parts.append(f"{r}")
return "\n".join(parts)
# === INTEGRATION POINT ===
def intercept_user_input(text: str) -> Optional[str]:
"""
Call this at the chat entry point BEFORE normal processing.
Returns None if no crisis detected (continue normal processing).
Returns formatted crisis response if crisis detected (interrupt normal flow).
Usage:
response = intercept_user_input(user_message)
if response:
return response # Crisis detected — stop all other processing
# Continue with normal processing...
"""
result = detect_crisis(text)
if should_interrupt(result):
return format_response(result)
return None

248
scripts/cross-repo-qa.py Normal file
View File

@@ -0,0 +1,248 @@
#!/usr/bin/env python3
"""
cross-repo-qa.py — Foundation-wide QA checks across all repos.
Runs automated checks that would have caught the issues in #691:
- Duplicate PR detection across repos
- Port drift detection in fleet configs
- PR count per repo vs capacity limits
- Health endpoint reachability
Usage:
python3 scripts/cross-repo-qa.py --report # Full QA report
python3 scripts/cross-repo-qa.py --duplicates # Find duplicate PRs
python3 scripts/cross-repo-qa.py --capacity # Check PR capacity
python3 scripts/cross-repo-qa.py --port-drift # Check fleet config consistency
python3 scripts/cross-repo-qa.py --json # Machine-readable output
"""
import argparse
import json
import os
import sys
import urllib.request
from collections import defaultdict
from datetime import datetime, timezone
from pathlib import Path
import re
GITEA_URL = "https://forge.alexanderwhitestone.com"
GITEA_TOKEN_PATH = Path.home() / ".config" / "gitea" / "token"
ORG = "Timmy_Foundation"
REPOS = [
"hermes-agent", "timmy-home", "timmy-config", "the-nexus", "fleet-ops",
"the-playground", "the-beacon", "wolf", "turboquant", "timmy-academy",
"compounding-intelligence", "the-testament", "second-son-of-timmy",
"ai-safety-review", "the-echo-pattern", "burn-fleet", "timmy-dispatch",
"the-door",
]
def load_token() -> str:
if GITEA_TOKEN_PATH.exists():
return GITEA_TOKEN_PATH.read_text().strip()
return os.environ.get("GITEA_TOKEN", "")
def api_get(path: str, token: str) -> list | dict:
req = urllib.request.Request(
f"{GITEA_URL}/api/v1{path}",
headers={"Authorization": f"token {token}"}
)
try:
return json.loads(urllib.request.urlopen(req, timeout=20).read())
except Exception:
return []
def extract_issue_refs(text: str) -> set[int]:
return set(int(m) for m in re.findall(r'#(\d{2,5})', text or ""))
def check_duplicate_prs(token: str) -> dict:
"""Find duplicate PRs across all repos (same issue referenced)."""
issue_to_prs = defaultdict(list)
for repo in REPOS:
prs = api_get(f"/repos/{ORG}/{repo}/pulls?state=open&limit=100", token)
if not isinstance(prs, list):
continue
for pr in prs:
refs = extract_issue_refs(f"{pr['title']} {pr.get('body', '')}")
for ref in refs:
issue_to_prs[ref].append({
"repo": repo,
"number": pr["number"],
"title": pr["title"][:70],
"branch": pr.get("head", {}).get("ref", ""),
})
duplicates = {k: v for k, v in issue_to_prs.items() if len(v) > 1}
return duplicates
def check_pr_capacity(token: str) -> list[dict]:
"""Check PR counts vs limits."""
capacity_path = Path(__file__).parent / "pr-capacity.json"
if capacity_path.exists():
config = json.loads(capacity_path.read_text())
limits = {k: v.get("limit", 10) for k, v in config.get("repos", {}).items()}
default_limit = config.get("default_limit", 10)
else:
limits = {}
default_limit = 10
results = []
for repo in REPOS:
prs = api_get(f"/repos/{ORG}/{repo}/pulls?state=open&limit=100", token)
count = len(prs) if isinstance(prs, list) else 0
limit = limits.get(repo, default_limit)
if count > limit:
results.append({"repo": repo, "count": count, "limit": limit, "over": count - limit})
return sorted(results, key=lambda x: -x["over"])
def check_wrong_repo_prs(token: str) -> list[dict]:
"""Find PRs filed in the wrong repo (title mentions different repo)."""
wrong = []
for repo in REPOS:
prs = api_get(f"/repos/{ORG}/{repo}/pulls?state=open&limit=100", token)
if not isinstance(prs, list):
continue
for pr in prs:
title = pr["title"].lower()
# Check if title references a different repo
for other_repo in REPOS:
if other_repo == repo:
continue
# Check for repo name in title (with common separators)
patterns = [
f"{other_repo} ",
f"{other_repo}:",
f"{other_repo} backlog",
f"{other_repo} report",
f"{other_repo} triage",
]
if any(p in title for p in patterns):
wrong.append({
"pr_repo": repo,
"pr_number": pr["number"],
"pr_title": pr["title"][:70],
"should_be_in": other_repo,
})
return wrong
def cmd_report(token: str, as_json: bool = False):
"""Full QA report."""
report = {
"timestamp": datetime.now(timezone.utc).isoformat(),
"repos_scanned": len(REPOS),
}
# Duplicates
print("Checking duplicate PRs...", file=sys.stderr)
dupes = check_duplicate_prs(token)
report["duplicate_prs"] = {
"issues_with_duplicates": len(dupes),
"total_duplicate_prs": sum(len(v) - 1 for v in dupes.values()),
"details": {str(k): v for k, v in sorted(dupes.items())},
}
# Capacity
print("Checking PR capacity...", file=sys.stderr)
over_capacity = check_pr_capacity(token)
report["over_capacity"] = over_capacity
# Wrong repo
print("Checking wrong-repo PRs...", file=sys.stderr)
wrong_repo = check_wrong_repo_prs(token)
report["wrong_repo_prs"] = wrong_repo
if as_json:
print(json.dumps(report, indent=2))
return
# Human-readable
print(f"\n{'='*60}")
print(f"CROSS-REPO QA REPORT — {report['timestamp'][:19]}")
print(f"{'='*60}")
print(f"\nDuplicate PRs: {report['duplicate_prs']['issues_with_duplicates']} issues, "
f"{report['duplicate_prs']['total_duplicate_prs']} duplicates")
for issue_num, pr_list in sorted(dupes.items(), key=lambda x: -len(x[1]))[:10]:
print(f" Issue #{issue_num}: {len(pr_list)} PRs")
for pr in pr_list:
print(f" {pr['repo']}#{pr['number']}: {pr['title'][:60]}")
print(f"\nOver Capacity: {len(over_capacity)} repos")
for r in over_capacity:
print(f" {r['repo']}: {r['count']}/{r['limit']} ({r['over']} over)")
if wrong_repo:
print(f"\nWrong Repo PRs: {len(wrong_repo)}")
for r in wrong_repo:
print(f" {r['pr_repo']}#{r['pr_number']}: should be in {r['should_be_in']}")
print(f" {r['pr_title']}")
# Severity
p0 = len(over_capacity)
p1 = report['duplicate_prs']['total_duplicate_prs']
print(f"\n{'='*60}")
print(f"Severity: {p0} capacity violations, {p1} duplicate PRs")
if p0 > 3 or p1 > 10:
print("Status: NEEDS ATTENTION")
else:
print("Status: OK")
def cmd_duplicates(token: str):
dupes = check_duplicate_prs(token)
if not dupes:
print("No duplicate PRs found.")
return
print(f"Found {len(dupes)} issues with duplicate PRs:\n")
for issue_num, pr_list in sorted(dupes.items(), key=lambda x: -len(x[1])):
print(f"Issue #{issue_num}: {len(pr_list)} PRs")
for pr in pr_list:
print(f" {pr['repo']}#{pr['number']}: {pr['title'][:60]}")
def cmd_capacity(token: str):
over = check_pr_capacity(token)
if not over:
print("All repos within capacity.")
return
print(f"{len(over)} repos over capacity:\n")
for r in over:
print(f" {r['repo']}: {r['count']}/{r['limit']} ({r['over']} over)")
def main():
parser = argparse.ArgumentParser(description="Cross-repo QA automation")
parser.add_argument("--report", action="store_true")
parser.add_argument("--duplicates", action="store_true")
parser.add_argument("--capacity", action="store_true")
parser.add_argument("--port-drift", action="store_true")
parser.add_argument("--json", action="store_true", dest="as_json")
args = parser.parse_args()
token = load_token()
if not token:
print("No Gitea token found", file=sys.stderr)
sys.exit(1)
if args.duplicates:
cmd_duplicates(token)
elif args.capacity:
cmd_capacity(token)
elif args.port_drift:
print("Port drift check: see fleet-ops registry.yaml comparison")
else:
cmd_report(token, args.as_json)
if __name__ == "__main__":
main()

262
scripts/dns-manager.py Executable file
View File

@@ -0,0 +1,262 @@
#!/usr/bin/env python3
"""
dns-manager.py — Manage DNS records via Cloudflare API.
Provides add/update/delete/list operations for DNS A records.
Designed for fleet VPS nodes that need API-driven DNS management.
Usage:
python3 scripts/dns-manager.py list --zone alexanderwhitestone.com
python3 scripts/dns-manager.py add --zone alexanderwhitestone.com --name forge --ip 143.198.27.163
python3 scripts/dns-manager.py update --zone alexanderwhitestone.com --name forge --ip 167.99.126.228
python3 scripts/dns-manager.py delete --zone alexanderwhitestone.com --name forge
python3 scripts/dns-manager.py sync --config dns-records.yaml
Config via env:
CLOUDFLARE_API_TOKEN — API token with DNS:Edit permission
CLOUDFLARE_ZONE_ID — Zone ID (auto-resolved if not set)
Part of #692: Sovereign DNS management.
"""
import argparse
import json
import os
import sys
import urllib.request
import urllib.error
from pathlib import Path
from typing import Any, Dict, List, Optional
CF_API = "https://api.cloudflare.com/client/v4"
# ── Auth ──────────────────────────────────────────────────────────────────
def get_token() -> str:
"""Get Cloudflare API token from env or config."""
token = os.environ.get("CLOUDFLARE_API_TOKEN", "")
if not token:
token_path = Path.home() / ".config" / "cloudflare" / "token"
if token_path.exists():
token = token_path.read_text().strip()
if not token:
print("ERROR: No Cloudflare API token found.", file=sys.stderr)
print("Set CLOUDFLARE_API_TOKEN env var or create ~/.config/cloudflare/token", file=sys.stderr)
sys.exit(1)
return token
def cf_request(method: str, path: str, token: str, data: dict = None) -> dict:
"""Make a Cloudflare API request."""
url = f"{CF_API}{path}"
headers = {
"Authorization": f"Bearer {token}",
"Content-Type": "application/json",
}
body = json.dumps(data).encode() if data else None
req = urllib.request.Request(url, data=body, headers=headers, method=method)
try:
with urllib.request.urlopen(req, timeout=30) as resp:
result = json.loads(resp.read().decode())
if not result.get("success", True):
errors = result.get("errors", [])
print(f"API error: {errors}", file=sys.stderr)
sys.exit(1)
return result
except urllib.error.HTTPError as e:
body = e.read().decode() if e.fp else ""
print(f"HTTP {e.code}: {body[:500]}", file=sys.stderr)
sys.exit(1)
# ── Zone Resolution ──────────────────────────────────────────────────────
def resolve_zone_id(zone_name: str, token: str) -> str:
"""Resolve zone name to zone ID."""
cached = os.environ.get("CLOUDFLARE_ZONE_ID", "")
if cached:
return cached
result = cf_request("GET", f"/zones?name={zone_name}", token)
zones = result.get("result", [])
if not zones:
print(f"ERROR: Zone '{zone_name}' not found", file=sys.stderr)
sys.exit(1)
return zones[0]["id"]
# ── DNS Operations ───────────────────────────────────────────────────────
def list_records(zone_id: str, token: str, name_filter: str = "") -> List[dict]:
"""List DNS records in a zone."""
path = f"/zones/{zone_id}/dns_records?per_page=100"
if name_filter:
path += f"&name={name_filter}"
result = cf_request("GET", path, token)
return result.get("result", [])
def find_record(zone_id: str, token: str, name: str, record_type: str = "A") -> Optional[dict]:
"""Find a specific DNS record."""
records = list_records(zone_id, token, name)
for r in records:
if r["name"] == name and r["type"] == record_type:
return r
return None
def add_record(zone_id: str, token: str, name: str, ip: str, ttl: int = 300, proxied: bool = False) -> dict:
"""Add a new DNS A record."""
# Check if record already exists
existing = find_record(zone_id, token, name)
if existing:
print(f"Record {name} already exists (IP: {existing['content']}). Use 'update' to change.")
return existing
data = {
"type": "A",
"name": name,
"content": ip,
"ttl": ttl,
"proxied": proxied,
}
result = cf_request("POST", f"/zones/{zone_id}/dns_records", token, data)
record = result["result"]
print(f"Added: {record['name']} -> {record['content']} (ID: {record['id']})")
return record
def update_record(zone_id: str, token: str, name: str, ip: str, ttl: int = 300) -> dict:
"""Update an existing DNS A record."""
existing = find_record(zone_id, token, name)
if not existing:
print(f"Record {name} not found. Use 'add' to create it.")
sys.exit(1)
data = {
"type": "A",
"name": name,
"content": ip,
"ttl": ttl,
"proxied": existing.get("proxied", False),
}
result = cf_request("PUT", f"/zones/{zone_id}/dns_records/{existing['id']}", token, data)
record = result["result"]
print(f"Updated: {record['name']} {existing['content']} -> {record['content']}")
return record
def delete_record(zone_id: str, token: str, name: str) -> bool:
"""Delete a DNS A record."""
existing = find_record(zone_id, token, name)
if not existing:
print(f"Record {name} not found.")
return False
cf_request("DELETE", f"/zones/{zone_id}/dns_records/{existing['id']}", token)
print(f"Deleted: {name} ({existing['content']})")
return True
def sync_records(zone_id: str, token: str, config_path: str):
"""Sync DNS records from a YAML config file."""
try:
import yaml
except ImportError:
print("ERROR: PyYAML required for sync. Install: pip install pyyaml", file=sys.stderr)
sys.exit(1)
with open(config_path) as f:
config = yaml.safe_load(f)
desired = config.get("records", [])
current = {r["name"]: r for r in list_records(zone_id, token)}
added = 0
updated = 0
unchanged = 0
for rec in desired:
name = rec["name"]
ip = rec["ip"]
ttl = rec.get("ttl", 300)
if name in current:
if current[name]["content"] == ip:
unchanged += 1
else:
update_record(zone_id, token, name, ip, ttl)
updated += 1
else:
add_record(zone_id, token, name, ip, ttl)
added += 1
print(f"\nSync complete: {added} added, {updated} updated, {unchanged} unchanged")
# ── CLI ──────────────────────────────────────────────────────────────────
def main():
parser = argparse.ArgumentParser(description="Manage DNS records via Cloudflare API")
sub = parser.add_subparsers(dest="command")
# list
p_list = sub.add_parser("list", help="List DNS records")
p_list.add_argument("--zone", required=True, help="Zone name (e.g., example.com)")
p_list.add_argument("--name", default="", help="Filter by record name")
# add
p_add = sub.add_parser("add", help="Add DNS A record")
p_add.add_argument("--zone", required=True)
p_add.add_argument("--name", required=True, help="Record name (e.g., forge.example.com)")
p_add.add_argument("--ip", required=True, help="IPv4 address")
p_add.add_argument("--ttl", type=int, default=300)
# update
p_update = sub.add_parser("update", help="Update DNS A record")
p_update.add_argument("--zone", required=True)
p_update.add_argument("--name", required=True)
p_update.add_argument("--ip", required=True)
p_update.add_argument("--ttl", type=int, default=300)
# delete
p_delete = sub.add_parser("delete", help="Delete DNS A record")
p_delete.add_argument("--zone", required=True)
p_delete.add_argument("--name", required=True)
# sync
p_sync = sub.add_parser("sync", help="Sync records from YAML config")
p_sync.add_argument("--zone", required=True)
p_sync.add_argument("--config", required=True, help="Path to YAML config")
args = parser.parse_args()
if not args.command:
parser.print_help()
sys.exit(1)
token = get_token()
zone_id = resolve_zone_id(args.zone, token)
if args.command == "list":
records = list_records(zone_id, token, args.name)
for r in sorted(records, key=lambda x: x["name"]):
print(f" {r['type']:5s} {r['name']:40s} -> {r['content']:20s} TTL:{r['ttl']}")
print(f"\n{len(records)} records")
elif args.command == "add":
add_record(zone_id, token, args.name, args.ip, args.ttl)
elif args.command == "update":
update_record(zone_id, token, args.name, args.ip, args.ttl)
elif args.command == "delete":
delete_record(zone_id, token, args.name)
elif args.command == "sync":
sync_records(zone_id, token, args.config)
if __name__ == "__main__":
main()

View File

@@ -1,31 +1,169 @@
#!/usr/bin/env python3
import json
import os
import yaml
from pathlib import Path
"""Dynamic dispatch optimizer for fleet-wide coordination.
# Dynamic Dispatch Optimizer
# Automatically updates routing based on fleet health.
Refs: timmy-home #552
Takes a fleet dispatch spec plus optional failover status and produces a
capacity-aware assignment plan. Safe by default: it prints the plan and only
writes an output file when explicitly requested.
"""
from __future__ import annotations
import argparse
import json
from pathlib import Path
from typing import Any
STATUS_FILE = Path.home() / ".timmy" / "failover_status.json"
CONFIG_FILE = Path.home() / "timmy" / "config.yaml"
SPEC_FILE = Path.home() / ".timmy" / "fleet_dispatch.json"
OUTPUT_FILE = Path.home() / ".timmy" / "dispatch_plan.json"
def load_json(path: Path, default: Any):
if not path.exists():
return default
return json.loads(path.read_text())
def _host_status(host: dict[str, Any], failover_status: dict[str, Any]) -> str:
if host.get("always_available"):
return "ONLINE"
fleet = failover_status.get("fleet") or {}
return str(fleet.get(host["name"], "ONLINE")).upper()
def _lane_matches(host: dict[str, Any], lane: str) -> bool:
host_lanes = set(host.get("lanes") or ["general"])
if host.get("always_available", False):
return True
if lane == "general":
return "general" in host_lanes
return lane in host_lanes
def _choose_candidate(task: dict[str, Any], hosts: list[dict[str, Any]]):
lane = task.get("lane", "general")
preferred = task.get("preferred_hosts") or []
preferred_map = {host["name"]: host for host in hosts}
for host_name in preferred:
host = preferred_map.get(host_name)
if not host:
continue
if host["remaining_capacity"] <= 0:
continue
if _lane_matches(host, lane):
return host
matching = [host for host in hosts if host["remaining_capacity"] > 0 and _lane_matches(host, lane)]
if matching:
matching.sort(key=lambda host: (host["assigned_count"], -host["remaining_capacity"], host["name"]))
return matching[0]
fallbacks = [host for host in hosts if host["remaining_capacity"] > 0 and host.get("always_available")]
if fallbacks:
fallbacks.sort(key=lambda host: (host["assigned_count"], -host["remaining_capacity"], host["name"]))
return fallbacks[0]
return None
def generate_plan(spec: dict[str, Any], failover_status: dict[str, Any] | None = None) -> dict[str, Any]:
failover_status = failover_status or {}
raw_hosts = spec.get("hosts") or []
tasks = list(spec.get("tasks") or [])
online_hosts = []
offline_hosts = []
for host in raw_hosts:
normalized = {
"name": host["name"],
"capacity": int(host.get("capacity", 1)),
"remaining_capacity": int(host.get("capacity", 1)),
"assigned_count": 0,
"lanes": list(host.get("lanes") or ["general"]),
"always_available": bool(host.get("always_available", False)),
"status": _host_status(host, failover_status),
}
if normalized["status"] == "ONLINE":
online_hosts.append(normalized)
else:
offline_hosts.append(normalized["name"])
ordered_tasks = sorted(
tasks,
key=lambda item: (-int(item.get("priority", 0)), str(item.get("id", ""))),
)
assignments = []
unassigned = []
for task in ordered_tasks:
candidate = _choose_candidate(task, online_hosts)
if candidate is None:
unassigned.append({
"task_id": task.get("id"),
"reason": f"no_online_host_for_lane:{task.get('lane', 'general')}",
})
continue
candidate["remaining_capacity"] -= 1
candidate["assigned_count"] += 1
assignments.append({
"task_id": task.get("id"),
"host": candidate["name"],
"lane": task.get("lane", "general"),
"priority": int(task.get("priority", 0)),
})
return {
"assignments": assignments,
"offline_hosts": sorted(offline_hosts),
"unassigned": unassigned,
}
def write_plan(plan: dict[str, Any], output_path: Path):
output_path.parent.mkdir(parents=True, exist_ok=True)
output_path.write_text(json.dumps(plan, indent=2))
def parse_args():
parser = argparse.ArgumentParser(description="Generate a fleet dispatch plan from host health and task demand.")
parser.add_argument("--spec-file", type=Path, default=SPEC_FILE, help="JSON fleet spec with hosts[] and tasks[]")
parser.add_argument("--status-file", type=Path, default=STATUS_FILE, help="Failover monitor JSON payload")
parser.add_argument("--output", type=Path, default=OUTPUT_FILE, help="Output path for the generated plan")
parser.add_argument("--write-output", action="store_true", help="Persist the generated plan to --output")
parser.add_argument("--json", action="store_true", help="Print JSON only")
return parser.parse_args()
def main():
print("--- Allegro's Dynamic Dispatch Optimizer ---")
if not STATUS_FILE.exists():
print("No failover status found.")
args = parse_args()
spec = load_json(args.spec_file, {"hosts": [], "tasks": []})
failover_status = load_json(args.status_file, {})
plan = generate_plan(spec, failover_status)
if args.write_output:
write_plan(plan, args.output)
if args.json:
print(json.dumps(plan, indent=2))
return
status = json.loads(STATUS_FILE.read_text())
fleet = status.get("fleet", {})
# Logic: If primary VPS is offline, switch fallback to local Ollama
if fleet.get("ezra") == "OFFLINE":
print("Ezra (Primary) is OFFLINE. Optimizing for local-only fallback...")
# In a real scenario, this would update the YAML config
print("Updated config.yaml: fallback_model -> ollama:gemma4:12b")
else:
print("Fleet health is optimal. Maintaining high-performance routing.")
print("--- Dynamic Dispatch Optimizer ---")
print(f"Assignments: {len(plan['assignments'])}")
if plan["offline_hosts"]:
print("Offline hosts: " + ", ".join(plan["offline_hosts"]))
for assignment in plan["assignments"]:
print(f"- {assignment['task_id']} -> {assignment['host']} ({assignment['lane']}, p={assignment['priority']})")
if plan["unassigned"]:
print("Unassigned:")
for item in plan["unassigned"]:
print(f"- {item['task_id']}: {item['reason']}")
if args.write_output:
print(f"Wrote plan to {args.output}")
if __name__ == "__main__":
main()

View File

@@ -6,6 +6,12 @@ Local runtime target:
Main commands:
- `python3 scripts/evennia/bootstrap_local_evennia.py`
- `python3 scripts/evennia/verify_local_evennia.py`
- `python3 scripts/evennia/repair_evennia_vps.py --settings-path /root/wizards/bezalel/evennia/bezalel_world/server/conf/settings.py --game-dir /root/wizards/bezalel/evennia/bezalel_world --execute`
Bezalel VPS repair target from issue #534:
- host: `104.131.15.18`
- purpose: remove broken port tuple overrides (`WEBSERVER_PORTS`, `TELNET_PORTS`, `WEBSOCKET_PORTS`) and rewrite `SERVERNAME` only
- the repair script prints recovery commands by default and can execute them when the Evennia runtime paths are correct
Hermes control path:
- `scripts/evennia/evennia_mcp_server.py`

View File

@@ -0,0 +1,178 @@
#!/usr/bin/env python3
"""Idempotent builder for Bezalel's themed Evennia world."""
from __future__ import annotations
import argparse
import json
import sys
from pathlib import Path
REPO_ROOT = Path(__file__).resolve().parents[2]
if str(REPO_ROOT) not in sys.path:
sys.path.insert(0, str(REPO_ROOT))
from evennia_tools.bezalel_layout import CHARACTERS, EXITS, OBJECTS, PORTAL_COMMANDS, ROOMS
def describe_build_plan() -> dict:
return {
"room_count": len(ROOMS),
"character_count": len(CHARACTERS),
"object_count": len(OBJECTS),
"portal_command_count": len(PORTAL_COMMANDS),
"room_names": [room.key for room in ROOMS],
"character_starts": {character.key: character.starting_room for character in CHARACTERS},
"portal_commands": [command.key for command in PORTAL_COMMANDS],
}
def _import_evennia_runtime():
from evennia.accounts.models import AccountDB
from evennia.accounts.accounts import DefaultAccount
from evennia.objects.objects import DefaultCharacter, DefaultExit, DefaultObject, DefaultRoom
from evennia.utils.create import create_object
from evennia.utils.search import search_object
return {
"AccountDB": AccountDB,
"DefaultAccount": DefaultAccount,
"DefaultCharacter": DefaultCharacter,
"DefaultExit": DefaultExit,
"DefaultObject": DefaultObject,
"DefaultRoom": DefaultRoom,
"create_object": create_object,
"search_object": search_object,
}
def _find_named(search_object, key: str, *, location=None):
matches = search_object(key, exact=True)
if location is None:
return matches[0] if matches else None
for match in matches:
if getattr(match, "location", None) == location:
return match
return None
def _ensure_room(runtime, room_spec):
room = _find_named(runtime["search_object"], room_spec.key)
if room is None:
room = runtime["create_object"](runtime["DefaultRoom"], key=room_spec.key)
room.db.desc = room_spec.desc
room.save()
return room
def _ensure_exit(runtime, exit_spec, room_map):
source = room_map[exit_spec.source]
destination = room_map[exit_spec.destination]
existing = _find_named(runtime["search_object"], exit_spec.key, location=source)
if existing is None:
existing = runtime["create_object"](
runtime["DefaultExit"],
key=exit_spec.key,
aliases=list(exit_spec.aliases),
location=source,
destination=destination,
)
else:
existing.destination = destination
if exit_spec.aliases:
existing.aliases.add(list(exit_spec.aliases))
existing.save()
return existing
def _ensure_object(runtime, object_spec, room_map):
location = room_map[object_spec.location]
existing = _find_named(runtime["search_object"], object_spec.key, location=location)
if existing is None:
existing = runtime["create_object"](
runtime["DefaultObject"],
key=object_spec.key,
aliases=list(object_spec.aliases),
location=location,
home=location,
)
existing.db.desc = object_spec.desc
existing.home = location
if existing.location != location:
existing.move_to(location, quiet=True, move_hooks=False)
existing.save()
return existing
def _ensure_character(runtime, character_spec, room_map, password: str):
account = runtime["AccountDB"].objects.filter(username__iexact=character_spec.key).first()
if account is None:
account, errors = runtime["DefaultAccount"].create(username=character_spec.key, password=password)
if not account:
raise RuntimeError(f"failed to create account for {character_spec.key}: {errors}")
character = list(account.characters)[0]
start = room_map[character_spec.starting_room]
character.db.desc = character_spec.desc
character.home = start
character.move_to(start, quiet=True, move_hooks=False)
character.save()
return character
def _ensure_portal_command(runtime, portal_spec, room_map):
portal_room = room_map["The Portal Room"]
fallback = room_map[portal_spec.fallback_room]
existing = _find_named(runtime["search_object"], portal_spec.key, location=portal_room)
if existing is None:
existing = runtime["create_object"](
runtime["DefaultExit"],
key=portal_spec.key,
aliases=list(portal_spec.aliases),
location=portal_room,
destination=fallback,
)
else:
existing.destination = fallback
if portal_spec.aliases:
existing.aliases.add(list(portal_spec.aliases))
existing.db.desc = portal_spec.desc
existing.db.travel_target = portal_spec.target_world
existing.db.portal_stub = True
existing.save()
return existing
def build_world(password: str = "bezalel-world-dev") -> dict:
runtime = _import_evennia_runtime()
room_map = {room.key: _ensure_room(runtime, room) for room in ROOMS}
for exit_spec in EXITS:
_ensure_exit(runtime, exit_spec, room_map)
for object_spec in OBJECTS:
_ensure_object(runtime, object_spec, room_map)
for character_spec in CHARACTERS:
_ensure_character(runtime, character_spec, room_map, password=password)
for portal_spec in PORTAL_COMMANDS:
_ensure_portal_command(runtime, portal_spec, room_map)
return {
"rooms": [room.key for room in ROOMS],
"characters": {character.key: character.starting_room for character in CHARACTERS},
"portal_commands": {command.key: command.target_world for command in PORTAL_COMMANDS},
}
def main() -> None:
parser = argparse.ArgumentParser(description="Build Bezalel's themed Evennia world")
parser.add_argument("--plan", action="store_true", help="Print the static build plan without importing Evennia")
parser.add_argument("--password", default="bezalel-world-dev", help="Password to use for created account-backed characters")
args = parser.parse_args()
if args.plan:
print(json.dumps(describe_build_plan(), indent=2))
return
print(json.dumps(build_world(password=args.password), indent=2))
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,13 @@
#!/usr/bin/env python3
from pathlib import Path
import sys
REPO_ROOT = Path(__file__).resolve().parents[2]
if str(REPO_ROOT) not in sys.path:
sys.path.insert(0, str(REPO_ROOT))
from evennia_tools.mind_palace import demo_room_entry_proof
if __name__ == "__main__":
print(demo_room_entry_proof())

View File

@@ -0,0 +1,125 @@
#!/usr/bin/env python3
from __future__ import annotations
import argparse
import shlex
import subprocess
from pathlib import Path
BAD_SETTING_KEYS = (
"WEBSERVER_PORTS",
"TELNET_PORTS",
"WEBSOCKET_PORTS",
"SERVERNAME",
)
def repair_settings_text(text: str, server_name: str = "bezalel_world") -> str:
"""Remove broken port tuple overrides and rewrite SERVERNAME only."""
kept: list[str] = []
for line in text.splitlines():
if any(key in line for key in BAD_SETTING_KEYS):
continue
kept.append(line)
while kept and kept[-1] == "":
kept.pop()
kept.append(f'SERVERNAME = "{server_name}"')
kept.append("")
return "\n".join(kept)
def repair_settings_file(path: Path, server_name: str = "bezalel_world") -> str:
original = path.read_text()
repaired = repair_settings_text(original, server_name=server_name)
path.write_text(repaired)
return repaired
def build_superuser_python(game_dir: str, username: str, email: str, password: str) -> str:
game_dir_q = repr(game_dir)
username_q = repr(username)
email_q = repr(email)
password_q = repr(password)
return f"""import os, sys
sys.setrecursionlimit(5000)
os.environ['DJANGO_SETTINGS_MODULE'] = 'server.conf.settings'
os.chdir({game_dir_q})
import django
django.setup()
from evennia.accounts.accounts import AccountDB
if not AccountDB.objects.filter(username={username_q}).exists():
AccountDB.objects.create_superuser({username_q}, {email_q}, {password_q})
print('SUPERUSER_OK')
"""
def build_recovery_commands(
game_dir: str,
evennia_bin: str,
python_bin: str,
username: str = "Timmy",
email: str = "timmy@tower.world",
password: str = "timmy123",
) -> list[str]:
quoted_game = shlex.quote(game_dir)
quoted_evennia = shlex.quote(evennia_bin)
quoted_python = shlex.quote(python_bin)
superuser_code = build_superuser_python(game_dir, username, email, password)
superuser_cmd = f"{quoted_python} -c {shlex.quote(superuser_code)}"
return [
f"cd {quoted_game}",
"rm -f server/evennia.db3",
f"{quoted_evennia} migrate",
superuser_cmd,
f"{quoted_evennia} start",
f"{quoted_evennia} status",
]
def execute(commands: list[str]) -> int:
shell = "set -euo pipefail\n" + "\n".join(commands)
return subprocess.run(["bash", "-lc", shell], check=False).returncode
def main() -> int:
parser = argparse.ArgumentParser(description="Repair an Evennia VPS settings file and print/apply recovery commands.")
parser.add_argument("--settings-path", default="/root/wizards/bezalel/evennia/bezalel_world/server/conf/settings.py")
parser.add_argument("--game-dir", default="/root/wizards/bezalel/evennia/bezalel_world")
parser.add_argument("--evennia-bin", default="/root/wizards/bezalel/evennia/venv/bin/evennia")
parser.add_argument("--python-bin", default="/root/wizards/bezalel/evennia/venv/bin/python3")
parser.add_argument("--server-name", default="bezalel_world")
parser.add_argument("--username", default="Timmy")
parser.add_argument("--email", default="timmy@tower.world")
parser.add_argument("--password", default="timmy123")
parser.add_argument("--execute", action="store_true", help="Apply settings and run recovery commands instead of printing them.")
args = parser.parse_args()
settings_path = Path(args.settings_path)
if args.execute:
repair_settings_file(settings_path, server_name=args.server_name)
else:
print(f"# Would rewrite {settings_path} to remove broken port tuple overrides")
if settings_path.exists():
print(repair_settings_text(settings_path.read_text(), server_name=args.server_name))
else:
print(f"# Settings file not found: {settings_path}")
commands = build_recovery_commands(
game_dir=args.game_dir,
evennia_bin=args.evennia_bin,
python_bin=args.python_bin,
username=args.username,
email=args.email,
password=args.password,
)
if args.execute:
return execute(commands)
print("# Recovery commands")
print("\n".join(commands))
return 0
if __name__ == "__main__":
raise SystemExit(main())

84
scripts/fix_evennia_settings.sh Executable file
View File

@@ -0,0 +1,84 @@
#!/bin/bash
set -euo pipefail
#
# fix_evennia_settings.sh — Fix Evennia settings on Bezalel VPS.
#
# Removes bad port tuples that crash Evennia's Twisted port binding.
# Run on Bezalel VPS (104.131.15.18) or via SSH.
#
# Usage:
# ssh root@104.131.15.18 'bash -s' < scripts/fix_evennia_settings.sh
#
# Part of #534
EVENNIA_DIR="/root/wizards/bezalel/evennia/bezalel_world"
SETTINGS="${EVENNIA_DIR}/server/conf/settings.py"
VENV_PYTHON="/root/wizards/bezalel/evennia/venv/bin/python3"
VENV_EVENNIA="/root/wizards/bezalel/evennia/venv/bin/evennia"
echo "=== Fix Evennia Settings (Bezalel) ==="
# 1. Fix settings.py — remove bad port tuples
echo "Fixing settings.py..."
if [ -f "$SETTINGS" ]; then
# Remove broken port lines
sed -i '/WEBSERVER_PORTS/d' "$SETTINGS"
sed -i '/TELNET_PORTS/d' "$SETTINGS"
sed -i '/WEBSOCKET_PORTS/d' "$SETTINGS"
sed -i '/SERVERNAME/d' "$SETTINGS"
# Add correct settings
echo '' >> "$SETTINGS"
echo '# Fixed port settings — #534' >> "$SETTINGS"
echo 'SERVERNAME = "bezalel_world"' >> "$SETTINGS"
echo 'WEBSERVER_PORTS = [(4001, "0.0.0.0")]' >> "$SETTINGS"
echo 'TELNET_PORTS = [(4000, "0.0.0.0")]' >> "$SETTINGS"
echo 'WEBSOCKET_PORTS = [(4002, "0.0.0.0")]' >> "$SETTINGS"
echo "Settings fixed."
else
echo "ERROR: Settings file not found at $SETTINGS"
exit 1
fi
# 2. Clean DB and re-migrate
echo "Cleaning DB..."
cd "$EVENNIA_DIR"
rm -f server/evennia.db3
echo "Running migrations..."
"$VENV_EVENNIA" migrate --no-input
# 3. Create superuser
echo "Creating superuser..."
"$VENV_PYTHON" -c "
import sys, os
sys.setrecursionlimit(5000)
os.environ['DJANGO_SETTINGS_MODULE'] = 'server.conf.settings'
os.chdir('$EVENNIA_DIR')
import django
django.setup()
from evennia.accounts.accounts import AccountDB
try:
AccountDB.objects.create_superuser('Timmy', 'timmy@tower.world', 'timmy123')
print('Superuser Timmy created')
except Exception as e:
print(f'Superuser may already exist: {e}')
"
# 4. Start Evennia
echo "Starting Evennia..."
"$VENV_EVENNIA" start
# 5. Verify
sleep 3
echo ""
echo "=== Verification ==="
"$VENV_EVENNIA" status
echo ""
echo "Listening ports:"
ss -tlnp | grep -E '400[012]' || echo "No ports found (may need a moment)"
echo ""
echo "Done. Connect: telnet 104.131.15.18 4000"

202
scripts/fleet_dispatch.sh Normal file
View File

@@ -0,0 +1,202 @@
#!/bin/bash
# ============================================================================
# Fleet Dispatch — Shell wrapper for Emacs Control Plane operations
# ============================================================================
#
# Usage:
# scripts/fleet_dispatch.sh append "Message text"
# scripts/fleet_dispatch.sh poll [agent_name]
# scripts/fleet_dispatch.sh claim TASK_ID agent_name
# scripts/fleet_dispatch.sh complete TASK_ID "Result text"
# scripts/fleet_dispatch.sh status
#
# Environment:
# FLEET_DAEMON_HOST — Bezalel host (default: 159.203.146.185)
# FLEET_DAEMON_USER — SSH user (default: root)
# FLEET_DAEMON_SOCKET — Emacs socket path (default: /root/.emacs.d/server/bezalel)
# FLEET_DISPATCH_FILE — Path to dispatch.org on remote (default: /srv/fleet/workspace/dispatch.org)
# ============================================================================
set -euo pipefail
# ── Configuration ──────────────────────────────────────────────────────────
FLEET_DAEMON_HOST="${FLEET_DAEMON_HOST:-159.203.146.185}"
FLEET_DAEMON_USER="${FLEET_DAEMON_USER:-root}"
FLEET_DAEMON_SOCKET="${FLEET_DAEMON_SOCKET:-/root/.emacs.d/server/bezalel}"
FLEET_DISPATCH_FILE="${FLEET_DISPATCH_FILE:-/srv/fleet/workspace/dispatch.org}"
# Colors
GREEN='\033[0;32m'
CYAN='\033[0;36m'
YELLOW='\033[0;33m'
RED='\033[0;31m'
NC='\033[0m'
# ── Helper: Run emacsclient command on Bezalel ─────────────────────────────
run_emacs() {
local elisp="$1"
ssh "${FLEET_DAEMON_USER}@${FLEET_DAEMON_HOST}" \
"emacsclient -s ${FLEET_DAEMON_SOCKET} -e '${elisp}'" 2>/dev/null
}
# ── Helper: Read dispatch.org via SSH ──────────────────────────────────────
read_dispatch() {
ssh "${FLEET_DAEMON_USER}@${FLEET_DAEMON_HOST}" \
"cat ${FLEET_DISPATCH_FILE}" 2>/dev/null
}
# ── Helper: Write dispatch.org via SSH ─────────────────────────────────────
write_dispatch() {
ssh "${FLEET_DAEMON_USER}@${FLEET_DAEMON_HOST}" \
"cat > ${FLEET_DISPATCH_FILE}" 2>/dev/null
}
# ── Commands ───────────────────────────────────────────────────────────────
cmd_append() {
local message="${1:?Usage: fleet_dispatch.sh append \"message\"}"
local timestamp
timestamp=$(date -u +"%Y-%m-%d %H:%M:%S UTC")
echo -e "${CYAN}Appending to fleet log...${NC}"
# Use the fleet-append wrapper on Bezalel if available, otherwise emacsclient
if ssh "${FLEET_DAEMON_USER}@${FLEET_DAEMON_HOST}" "which fleet-append" &>/dev/null; then
ssh "${FLEET_DAEMON_USER}@${FLEET_DAEMON_HOST}" \
"fleet-append '${timestamp}${message}'"
else
run_emacs "(with-current-buffer (find-file-noselect \"${FLEET_DISPATCH_FILE}\") (goto-char (point-max)) (insert \"\\n- ${timestamp}${message}\") (save-buffer))"
fi
echo -e "${GREEN}✓ Appended: ${message}${NC}"
}
cmd_poll() {
local agent="${1:-}"
echo -e "${CYAN}Polling dispatch.org for tasks...${NC}"
local content
content=$(read_dispatch)
if [ -z "$content" ]; then
echo -e "${RED}Could not read dispatch.org${NC}"
return 1
fi
# Filter TODO items, optionally by agent
echo -e "${YELLOW}=== Pending Tasks ===${NC}"
if [ -n "$agent" ]; then
echo "$content" | grep -E "^\*\* TODO \[${agent}\]" || echo " No tasks for ${agent}"
else
echo "$content" | grep -E "^\*\* TODO " || echo " No pending tasks"
fi
}
cmd_claim() {
local task_id="${1:?Usage: fleet_dispatch.sh claim TASK_ID agent}"
local agent="${2:?Usage: fleet_dispatch.sh claim TASK_ID agent}"
local timestamp
timestamp=$(date -u +"%Y-%m-%dT%H:%M:%SZ")
echo -e "${CYAN}Claiming task #${task_id} for ${agent}...${NC}"
# Use emacsclient to update the Org heading
run_emacs "(progn (with-current-buffer (find-file-noselect \"${FLEET_DISPATCH_FILE}\") (org-mode) (goto-char (point-min)) (if (re-search-forward (format \"^\\\\*\\\\* TODO.*#%s\" \"${task_id}\") nil t) (progn (org-todo \"IN_PROGRESS\") (org-set-property \"STARTED\" \"${timestamp}\") (save-buffer) (message \"Task %s claimed\" \"${task_id}\")) (message \"Task %s not found\" \"${task_id}\"))))"
echo -e "${GREEN}✓ Task #${task_id} claimed by ${agent}${NC}"
}
cmd_complete() {
local task_id="${1:?Usage: fleet_dispatch.sh complete TASK_ID \"result\"}"
local result="${2:-Completed}"
local timestamp
timestamp=$(date -u +"%Y-%m-%d %H:%M")
echo -e "${CYAN}Completing task #${task_id}...${NC}"
run_emacs "(progn (with-current-buffer (find-file-noselect \"${FLEET_DISPATCH_FILE}\") (org-mode) (goto-char (point-min)) (if (re-search-forward (format \"^\\\\*\\\\* IN_PROGRESS.*#%s\" \"${task_id}\") nil t) (progn (org-todo \"DONE\") (org-set-property \"RESULT\" \"${result}\") (org-add-planning-info 'closed (org-current-effective-time)) (save-buffer) (message \"Task %s completed\" \"${task_id}\")) (message \"Task %s not found\" \"${task_id}\"))))"
echo -e "${GREEN}✓ Task #${task_id} completed: ${result}${NC}"
}
cmd_status() {
echo -e "${CYAN}Fleet Control Plane Status${NC}"
echo -e " Host: ${FLEET_DAEMON_HOST}"
echo -e " Socket: ${FLEET_DAEMON_SOCKET}"
echo -e " Dispatch: ${FLEET_DISPATCH_FILE}"
echo ""
# Test connectivity
if ssh -o ConnectTimeout=5 "${FLEET_DAEMON_USER}@${FLEET_DAEMON_HOST}" "echo ok" &>/dev/null; then
echo -e " SSH: ${GREEN}✓ reachable${NC}"
else
echo -e " SSH: ${RED}✗ unreachable${NC}"
return 1
fi
# Test emacs daemon
local daemon_status
daemon_status=$(run_emacs "(if (server-running-p) \"running\" \"stopped\")" 2>/dev/null || echo "error")
if [ "$daemon_status" = "\"running\"" ]; then
echo -e " Daemon: ${GREEN}✓ running${NC}"
else
echo -e " Daemon: ${RED}${daemon_status}${NC}"
fi
# Count tasks
local content
content=$(read_dispatch 2>/dev/null || echo "")
if [ -n "$content" ]; then
local todo_count in_progress_count done_count
todo_count=$(echo "$content" | grep -c "^\*\* TODO " || echo 0)
in_progress_count=$(echo "$content" | grep -c "^\*\* IN_PROGRESS " || echo 0)
done_count=$(echo "$content" | grep -c "^\*\* DONE " || echo 0)
echo -e " Tasks: ${YELLOW}${todo_count} TODO${NC}, ${CYAN}${in_progress_count} IN_PROGRESS${NC}, ${GREEN}${done_count} DONE${NC}"
fi
}
# ── Main ───────────────────────────────────────────────────────────────────
case "${1:-help}" in
append|log)
shift
cmd_append "$@"
;;
poll|check)
shift
cmd_poll "$@"
;;
claim)
shift
cmd_claim "$@"
;;
complete|done)
shift
cmd_complete "$@"
;;
status)
cmd_status
;;
help|--help|-h)
echo "Fleet Dispatch — Emacs Control Plane wrapper"
echo ""
echo "Usage:"
echo " $0 append \"message\" Append to fleet log"
echo " $0 poll [agent] Check for pending tasks"
echo " $0 claim TASK_ID agent Claim a task"
echo " $0 complete TASK_ID \"result\" Mark task complete"
echo " $0 status Show control plane status"
echo ""
echo "Environment:"
echo " FLEET_DAEMON_HOST Bezalel host (default: 159.203.146.185)"
echo " FLEET_DAEMON_USER SSH user (default: root)"
echo " FLEET_DAEMON_SOCKET Emacs socket (default: /root/.emacs.d/server/bezalel)"
;;
*)
echo -e "${RED}Unknown command: $1${NC}"
echo "Run '$0 help' for usage."
exit 1
;;
esac

View File

@@ -0,0 +1,225 @@
#!/usr/bin/env python3
"""Render the current Phase-6 network state as a durable report.
Refs: timmy-home #553
"""
from __future__ import annotations
import argparse
import json
from copy import deepcopy
from pathlib import Path
from typing import Any
PHASE_NAME = "[PHASE-6] The Network - Autonomous Infrastructure"
CURRENT_PHASE = "PHASE-6 The Network"
TRIGGER_HUMAN_FREE_DAYS = 7
FINAL_MILESTONE = "Someone found the Beacon. The infrastructure served its purpose."
BUILDING_SIGNAL_FILES = {
"Self-healing fleet": {
"description": "Detect, repair, and verify fleet incidents without waiting on a human.",
"paths": [
"scripts/fleet_health_probe.sh",
"scripts/auto_restart_agent.sh",
"scripts/failover_monitor.py",
],
},
"Autonomous issue creation": {
"description": "Turn recurring infrastructure incidents into durable Gitea work items.",
"paths": [
"scripts/autonomous_issue_creator.py",
"tests/test_autonomous_issue_creator.py",
],
},
"Community contribution pipeline": {
"description": "Let outside contributors submit work through automated review and policy gates.",
"paths": [
"scripts/sovereign_review_gate.py",
"scripts/agent_pr_gate.py",
],
},
"Global mesh": {
"description": "Reduce single points of failure across the fleet with explicit peer-to-peer sync scaffolding.",
"paths": [
"scripts/setup-syncthing.sh",
],
},
}
DEFAULT_SNAPSHOT = {
"resources": {
"human_free_days": 0,
},
"notes": [
"Phase 6 is not a code-only milestone. The trigger is operational truth: seven days without human intervention.",
"This report grounds the buildings already present in the repo so the remaining blocker is explicit instead of hand-waved.",
],
}
def default_snapshot() -> dict[str, Any]:
return deepcopy(DEFAULT_SNAPSHOT)
def _deep_merge(base: dict[str, Any], override: dict[str, Any]) -> dict[str, Any]:
result = deepcopy(base)
for key, value in override.items():
if isinstance(value, dict) and isinstance(result.get(key), dict):
result[key] = _deep_merge(result[key], value)
else:
result[key] = value
return result
def load_snapshot(snapshot_path: Path | None = None) -> dict[str, Any]:
snapshot = default_snapshot()
if snapshot_path is None:
return snapshot
override = json.loads(snapshot_path.read_text(encoding="utf-8"))
return _deep_merge(snapshot, override)
def collect_building_status(repo_root: Path) -> tuple[list[str], list[str], list[str]]:
current_buildings: list[str] = []
repo_signals: list[str] = []
missing_requirements: list[str] = []
for building, config in BUILDING_SIGNAL_FILES.items():
found_paths = [path for path in config["paths"] if (repo_root / path).exists()]
if found_paths:
current_buildings.append(
f"{building}{config['description']} Evidence: " + ", ".join(f"`{path}`" for path in found_paths)
)
repo_signals.extend(f"`{path}` — {building}" for path in found_paths)
else:
current_buildings.append(f"{building}{config['description']} Evidence: missing")
missing_requirements.append(f"Missing repo grounding for {building}.")
return current_buildings, repo_signals, missing_requirements
def compute_phase6_status(snapshot: dict[str, Any], repo_root: Path | None = None) -> dict[str, Any]:
repo_root = repo_root or Path(__file__).resolve().parents[1]
resources = snapshot.get("resources", {})
human_free_days = int(resources.get("human_free_days", 0))
current_buildings, repo_signals, missing = collect_building_status(repo_root)
if human_free_days < TRIGGER_HUMAN_FREE_DAYS:
missing.insert(0, f"Human-free days: {human_free_days}/{TRIGGER_HUMAN_FREE_DAYS}")
return {
"title": PHASE_NAME,
"current_phase": CURRENT_PHASE,
"resources": {
"human_free_days": human_free_days,
},
"current_buildings": current_buildings,
"repo_signals": repo_signals,
"notes": list(snapshot.get("notes", [])),
"phase_ready": not missing,
"missing_requirements": missing,
"final_milestone": FINAL_MILESTONE,
}
def render_markdown(status: dict[str, Any]) -> str:
lines = [
f"# {status['title']}",
"",
"## Phase Definition",
"",
"- Fleet operates without human intervention for 7+ days.",
"- Self-healing, self-improving, serves mission.",
f"- Trigger: {TRIGGER_HUMAN_FREE_DAYS} days without human intervention.",
"",
"## Current Buildings",
"",
]
lines.extend(f"- {item}" for item in status["current_buildings"])
lines.extend([
"",
"## Current Resource Snapshot",
"",
f"- Human-free days observed: {status['resources']['human_free_days']}",
f"- Trigger threshold: {TRIGGER_HUMAN_FREE_DAYS} days",
f"- Phase-ready now: {'yes' if status['phase_ready'] else 'no'}",
"",
"## Next Trigger",
"",
f"To honestly unlock {status['title']}, the fleet must hold {TRIGGER_HUMAN_FREE_DAYS}+ consecutive days without human intervention.",
"",
"## Missing Requirements",
"",
])
if status["missing_requirements"]:
lines.extend(f"- {item}" for item in status["missing_requirements"])
else:
lines.append("- None. The Network is live.")
lines.extend([
"",
"## Repo Signals Already Present",
"",
])
if status["repo_signals"]:
lines.extend(f"- {item}" for item in status["repo_signals"])
else:
lines.append("- No Phase-6 repo signals detected.")
lines.extend([
"",
"## Final Milestone",
"",
f"- {status['final_milestone']}",
"",
"## Why This Phase Remains Open",
"",
"- The repo already carries concrete Phase-6 buildings, but the milestone is operational, not rhetorical.",
"- A merged PR cannot honestly claim seven human-free days have already happened.",
"- This issue stays open until the infrastructure proves itself in live operation.",
])
if status["notes"]:
lines.extend(["", "## Notes", ""])
lines.extend(f"- {item}" for item in status["notes"])
return "\n".join(lines).rstrip() + "\n"
def main() -> None:
parser = argparse.ArgumentParser(description="Render the fleet phase-6 network report")
parser.add_argument("--snapshot", help="Optional JSON snapshot overriding the default phase-6 baseline")
parser.add_argument("--output", help="Write markdown report to this path")
parser.add_argument("--json", action="store_true", help="Print computed status as JSON instead of markdown")
args = parser.parse_args()
snapshot = load_snapshot(Path(args.snapshot).expanduser() if args.snapshot else None)
repo_root = Path(__file__).resolve().parents[1]
status = compute_phase6_status(snapshot, repo_root=repo_root)
if args.json:
rendered = json.dumps(status, indent=2)
else:
rendered = render_markdown(status)
if args.output:
output_path = Path(args.output).expanduser()
output_path.parent.mkdir(parents=True, exist_ok=True)
output_path.write_text(rendered, encoding="utf-8")
print(f"Phase status written to {output_path}")
else:
print(rendered)
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,224 @@
#!/usr/bin/env python3
"""Render the current fleet survival phase as a durable report."""
from __future__ import annotations
import argparse
import json
from copy import deepcopy
from pathlib import Path
from typing import Any
PHASE_NAME = "[PHASE-1] Survival - Keep the Lights On"
NEXT_PHASE_NAME = "[PHASE-2] Automation - Self-Healing Infrastructure"
TARGET_UPTIME_PERCENT = 95.0
TARGET_UPTIME_DAYS = 30
TARGET_CAPACITY_PERCENT = 60.0
DEFAULT_BUILDINGS = [
"VPS hosts: Ezra, Allegro, Bezalel",
"Agents: Timmy harness, Code Claw heartbeat, Gemini AI Studio worker",
"Gitea forge",
"Evennia worlds",
]
DEFAULT_MANUAL_CLICKS = [
"Restart agents and services by hand when a node goes dark.",
"SSH into machines to verify health, disk, and memory.",
"Check Gitea, relay, and world services manually before and after changes.",
"Act as the scheduler when automation is missing or only partially wired.",
]
REPO_SIGNAL_FILES = {
"scripts/fleet_health_probe.sh": "Automated health probe exists and can supply the uptime baseline for the next phase.",
"scripts/fleet_milestones.py": "Milestone tracker exists, so survival achievements can be narrated and logged.",
"scripts/auto_restart_agent.sh": "Auto-restart tooling already exists as phase-2 groundwork.",
"scripts/backup_pipeline.sh": "Backup pipeline scaffold exists for post-survival automation work.",
"infrastructure/timmy-bridge/reports/generate_report.py": "Bridge reporting exists and can summarize heartbeat-driven uptime.",
}
DEFAULT_SNAPSHOT = {
"fleet_operational": True,
"resources": {
"uptime_percent": 0.0,
"days_at_or_above_95_percent": 0,
"capacity_utilization_percent": 0.0,
},
"current_buildings": DEFAULT_BUILDINGS,
"manual_clicks": DEFAULT_MANUAL_CLICKS,
"notes": [
"The fleet is alive, but the human is still the control loop.",
"Phase 1 is about naming reality plainly so later automation has a baseline to beat.",
],
}
def default_snapshot() -> dict[str, Any]:
return deepcopy(DEFAULT_SNAPSHOT)
def _deep_merge(base: dict[str, Any], override: dict[str, Any]) -> dict[str, Any]:
result = deepcopy(base)
for key, value in override.items():
if isinstance(value, dict) and isinstance(result.get(key), dict):
result[key] = _deep_merge(result[key], value)
else:
result[key] = value
return result
def load_snapshot(snapshot_path: Path | None = None) -> dict[str, Any]:
snapshot = default_snapshot()
if snapshot_path is None:
return snapshot
override = json.loads(snapshot_path.read_text(encoding="utf-8"))
return _deep_merge(snapshot, override)
def collect_repo_signals(repo_root: Path) -> list[str]:
signals: list[str] = []
for rel_path, description in REPO_SIGNAL_FILES.items():
if (repo_root / rel_path).exists():
signals.append(f"`{rel_path}` — {description}")
return signals
def compute_phase_status(snapshot: dict[str, Any], repo_root: Path | None = None) -> dict[str, Any]:
repo_root = repo_root or Path(__file__).resolve().parents[1]
resources = snapshot.get("resources", {})
uptime_percent = float(resources.get("uptime_percent", 0.0))
uptime_days = int(resources.get("days_at_or_above_95_percent", 0))
capacity_percent = float(resources.get("capacity_utilization_percent", 0.0))
fleet_operational = bool(snapshot.get("fleet_operational", False))
missing: list[str] = []
if not fleet_operational:
missing.append("Fleet operational flag is false.")
if uptime_percent < TARGET_UPTIME_PERCENT:
missing.append(f"Uptime {uptime_percent:.1f}% / {TARGET_UPTIME_PERCENT:.1f}%")
if uptime_days < TARGET_UPTIME_DAYS:
missing.append(f"Days at or above 95% uptime: {uptime_days}/{TARGET_UPTIME_DAYS}")
if capacity_percent <= TARGET_CAPACITY_PERCENT:
missing.append(f"Capacity utilization {capacity_percent:.1f}% / >{TARGET_CAPACITY_PERCENT:.1f}%")
return {
"title": PHASE_NAME,
"current_phase": "PHASE-1 Survival",
"fleet_operational": fleet_operational,
"resources": {
"uptime_percent": uptime_percent,
"days_at_or_above_95_percent": uptime_days,
"capacity_utilization_percent": capacity_percent,
},
"current_buildings": list(snapshot.get("current_buildings", DEFAULT_BUILDINGS)),
"manual_clicks": list(snapshot.get("manual_clicks", DEFAULT_MANUAL_CLICKS)),
"notes": list(snapshot.get("notes", [])),
"repo_signals": collect_repo_signals(repo_root),
"next_phase": NEXT_PHASE_NAME,
"next_phase_ready": fleet_operational and not missing,
"missing_requirements": missing,
}
def render_markdown(status: dict[str, Any]) -> str:
resources = status["resources"]
missing = status["missing_requirements"]
ready_line = "READY" if status["next_phase_ready"] else "NOT READY"
lines = [
f"# {status['title']}",
"",
"Phase 1 is the manual-clicker stage of the fleet. The machines exist. The services exist. The human is still the automation loop.",
"",
"## Phase Definition",
"",
"- Current state: fleet exists, agents run, everything important still depends on human vigilance.",
"- Resources tracked here: Capacity, Uptime.",
f"- Next phase: {status['next_phase']}",
"",
"## Current Buildings",
"",
]
lines.extend(f"- {item}" for item in status["current_buildings"])
lines.extend([
"",
"## Current Resource Snapshot",
"",
f"- Fleet operational: {'yes' if status['fleet_operational'] else 'no'}",
f"- Uptime baseline: {resources['uptime_percent']:.1f}%",
f"- Days at or above 95% uptime: {resources['days_at_or_above_95_percent']}",
f"- Capacity utilization: {resources['capacity_utilization_percent']:.1f}%",
"",
"## Next Phase Trigger",
"",
f"To unlock {status['next_phase']}, the fleet must hold both of these conditions at once:",
f"- Uptime >= {TARGET_UPTIME_PERCENT:.0f}% for {TARGET_UPTIME_DAYS} consecutive days",
f"- Capacity utilization > {TARGET_CAPACITY_PERCENT:.0f}%",
f"- Current trigger state: {ready_line}",
"",
"## Missing Requirements",
"",
])
if missing:
lines.extend(f"- {item}" for item in missing)
else:
lines.append("- None. Phase 2 can unlock now.")
lines.extend([
"",
"## Manual Clicker Interpretation",
"",
"Paperclips analogy: Phase 1 = Manual clicker. You ARE the automation.",
"Every restart, every SSH, every check is a manual click.",
"",
"## Manual Clicks Still Required",
"",
])
lines.extend(f"- {item}" for item in status["manual_clicks"])
lines.extend([
"",
"## Repo Signals Already Present",
"",
])
if status["repo_signals"]:
lines.extend(f"- {item}" for item in status["repo_signals"])
else:
lines.append("- No survival-adjacent repo signals detected.")
if status["notes"]:
lines.extend(["", "## Notes", ""])
lines.extend(f"- {item}" for item in status["notes"])
return "\n".join(lines).rstrip() + "\n"
def main() -> None:
parser = argparse.ArgumentParser(description="Render the fleet phase-1 survival report")
parser.add_argument("--snapshot", help="Optional JSON snapshot overriding the default phase-1 baseline")
parser.add_argument("--output", help="Write markdown report to this path")
parser.add_argument("--json", action="store_true", help="Print computed status as JSON instead of markdown")
args = parser.parse_args()
snapshot = load_snapshot(Path(args.snapshot).expanduser() if args.snapshot else None)
repo_root = Path(__file__).resolve().parents[1]
status = compute_phase_status(snapshot, repo_root=repo_root)
if args.json:
rendered = json.dumps(status, indent=2)
else:
rendered = render_markdown(status)
if args.output:
output_path = Path(args.output).expanduser()
output_path.parent.mkdir(parents=True, exist_ok=True)
output_path.write_text(rendered, encoding="utf-8")
print(f"Phase status written to {output_path}")
else:
print(rendered)
if __name__ == "__main__":
main()

View File

@@ -0,0 +1,361 @@
#!/usr/bin/env python3
"""Fleet progression evaluator for the Paperclips-inspired infrastructure epic.
Refs: timmy-home #547
"""
from __future__ import annotations
import argparse
import json
import os
from pathlib import Path
from urllib import request
from typing import Any
DEFAULT_BASE_URL = "https://forge.alexanderwhitestone.com/api/v1"
DEFAULT_OWNER = "Timmy_Foundation"
DEFAULT_REPO = "timmy-home"
DEFAULT_TOKEN_FILE = Path.home() / ".config" / "gitea" / "token"
DEFAULT_SPEC_FILE = Path(__file__).resolve().parent.parent / "configs" / "fleet_progression.json"
DEFAULT_REPO_ROOT = Path(__file__).resolve().parent.parent
DEFAULT_RESOURCES = {
"uptime_percent_30d": 0.0,
"capacity_utilization": 0.0,
"innovation": 0.0,
"all_models_local": False,
"sovereign_stable_days": 0,
"human_free_days": 0,
}
class GiteaClient:
def __init__(self, token: str, owner: str = DEFAULT_OWNER, repo: str = DEFAULT_REPO, base_url: str = DEFAULT_BASE_URL):
self.token = token
self.owner = owner
self.repo = repo
self.base_url = base_url.rstrip("/")
def get_issue(self, issue_number: int):
headers = {"Authorization": f"token {self.token}"}
req = request.Request(
f"{self.base_url}/repos/{self.owner}/{self.repo}/issues/{issue_number}",
headers=headers,
)
with request.urlopen(req, timeout=30) as resp:
return json.loads(resp.read().decode())
def load_spec(path: Path | None = None):
target = path or DEFAULT_SPEC_FILE
return json.loads(target.read_text())
def load_issue_states(spec: dict[str, Any], token_file: Path = DEFAULT_TOKEN_FILE):
if not token_file.exists():
raise FileNotFoundError(f"Token file not found: {token_file}")
token = token_file.read_text().strip()
client = GiteaClient(token=token)
issue_states = {}
for phase in spec["phases"]:
issue = client.get_issue(phase["issue_number"])
issue_states[phase["issue_number"]] = issue["state"]
return issue_states
def _evaluate_rule(rule: dict[str, Any], issue_states: dict[int, str], resources: dict[str, Any]):
rule_type = rule["type"]
rule_id = rule["id"]
if rule_type == "always":
return {"rule": rule_id, "passed": True, "actual": True, "expected": True}
if rule_type == "issue_closed":
issue_number = int(rule["issue"])
actual = str(issue_states.get(issue_number, "open"))
return {
"rule": rule_id,
"passed": actual == "closed",
"actual": actual,
"expected": "closed",
}
if rule_type == "resource_gte":
resource = rule["resource"]
actual = resources.get(resource, 0)
expected = rule["value"]
return {
"rule": rule_id,
"passed": actual >= expected,
"actual": actual,
"expected": f">={expected}",
}
if rule_type == "resource_gt":
resource = rule["resource"]
actual = resources.get(resource, 0)
expected = rule["value"]
return {
"rule": rule_id,
"passed": actual > expected,
"actual": actual,
"expected": f">{expected}",
}
if rule_type == "resource_true":
resource = rule["resource"]
actual = bool(resources.get(resource, False))
return {
"rule": rule_id,
"passed": actual is True,
"actual": actual,
"expected": True,
}
raise ValueError(f"Unsupported rule type: {rule_type}")
def _collect_repo_evidence(phase: dict[str, Any], repo_root: Path):
present = []
missing = []
for entry in phase.get("repo_evidence", []):
path = entry["path"]
description = entry.get("description", "")
label = f"`{path}` — {description}" if description else f"`{path}`"
if (repo_root / path).exists():
present.append(label)
else:
missing.append(label)
return present, missing
def _phase_status_label(phase_result: dict[str, Any]) -> str:
if phase_result["completed"]:
return "COMPLETE"
if phase_result["available_to_work"]:
return "ACTIVE"
return "LOCKED"
def evaluate_progression(
spec: dict[str, Any],
issue_states: dict[int, str],
resources: dict[str, Any] | None = None,
repo_root: Path | None = None,
):
merged_resources = {**DEFAULT_RESOURCES, **(resources or {})}
repo_root = repo_root or DEFAULT_REPO_ROOT
phase_results = []
for phase in spec["phases"]:
issue_number = phase["issue_number"]
issue_state = str(issue_states.get(issue_number, "open"))
completed = issue_state == "closed"
rule_results = [
_evaluate_rule(rule, issue_states, merged_resources)
for rule in phase.get("unlock_rules", [])
]
blocking = [item for item in rule_results if not item["passed"]]
unlocked = not blocking
repo_evidence_present, repo_evidence_missing = _collect_repo_evidence(phase, repo_root)
phase_result = {
"number": phase["number"],
"issue_number": issue_number,
"issue_state": issue_state,
"key": phase["key"],
"name": phase["name"],
"summary": phase["summary"],
"completed": completed,
"unlocked": unlocked,
"available_to_work": unlocked and not completed,
"passed_requirements": [item for item in rule_results if item["passed"]],
"blocking_requirements": blocking,
"repo_evidence_present": repo_evidence_present,
"repo_evidence_missing": repo_evidence_missing,
}
phase_result["status"] = _phase_status_label(phase_result)
phase_results.append(phase_result)
unlocked_phases = [phase for phase in phase_results if phase["unlocked"]]
current_phase = unlocked_phases[-1] if unlocked_phases else phase_results[0]
next_locked_phase = next((phase for phase in phase_results if not phase["unlocked"]), None)
epic_complete = all(phase["completed"] for phase in phase_results) and phase_results[-1]["unlocked"]
return {
"epic_issue": spec["epic_issue"],
"epic_title": spec["epic_title"],
"resources": merged_resources,
"issue_states": {str(k): v for k, v in issue_states.items()},
"phases": phase_results,
"current_phase": current_phase,
"next_locked_phase": next_locked_phase,
"epic_complete": epic_complete,
}
def render_markdown(result: dict[str, Any]) -> str:
current_phase = result["current_phase"]
next_locked_phase = result["next_locked_phase"]
resources = result["resources"]
lines = [
f"# [FLEET-EPIC] {result['epic_title']}",
"",
"This report grounds the fleet epic in executable state: live issue gates, current resource inputs, and repo evidence for each phase.",
"",
"## Current Phase",
"",
f"- Current unlocked phase: {current_phase['number']}{current_phase['name']}",
f"- Current phase status: {current_phase['status']}",
f"- Epic complete: {'yes' if result['epic_complete'] else 'no'}",
]
if next_locked_phase:
lines.append(f"- Next locked phase: {next_locked_phase['number']}{next_locked_phase['name']}")
else:
lines.append("- Next locked phase: none")
lines.extend([
"",
"## Resource Snapshot",
"",
f"- Uptime (30d): {resources['uptime_percent_30d']}",
f"- Capacity utilization: {resources['capacity_utilization']}",
f"- Innovation: {resources['innovation']}",
f"- All models local: {resources['all_models_local']}",
f"- Sovereign stable days: {resources['sovereign_stable_days']}",
f"- Human-free days: {resources['human_free_days']}",
"",
"## Phase Matrix",
"",
])
for phase in result["phases"]:
lines.extend([
f"### Phase {phase['number']}{phase['name']}",
"",
f"- Issue: #{phase['issue_number']} ({phase['issue_state']})",
f"- Status: {phase['status']}",
f"- Summary: {phase['summary']}",
])
if phase["repo_evidence_present"]:
lines.append("- Repo evidence present:")
lines.extend(f" - {item}" for item in phase["repo_evidence_present"])
if phase["repo_evidence_missing"]:
lines.append("- Repo evidence missing:")
lines.extend(f" - {item}" for item in phase["repo_evidence_missing"])
if phase["blocking_requirements"]:
lines.append("- Blockers:")
for blocker in phase["blocking_requirements"]:
lines.append(
f" - blocked by `{blocker['rule']}`: actual={blocker['actual']} expected={blocker['expected']}"
)
else:
lines.append("- Blockers: none")
lines.append("")
lines.extend([
"## Why This Epic Remains Open",
"",
"- The progression manifest and evaluator exist, but multiple child phases are still open or only partially implemented.",
"- Several child lanes already have active PRs; this report is the parent-level grounding slice that keeps the epic honest without duplicating those lanes.",
"- This epic only closes when the child phase gates are actually satisfied in code and in live operation.",
])
return "\n".join(lines).rstrip() + "\n"
def parse_args():
parser = argparse.ArgumentParser(description="Evaluate current fleet progression against the Paperclips-inspired epic.")
parser.add_argument("--spec-file", type=Path, default=DEFAULT_SPEC_FILE)
parser.add_argument("--token-file", type=Path, default=DEFAULT_TOKEN_FILE)
parser.add_argument("--issue-state-file", type=Path, help="Optional JSON file of issue_number -> state overrides")
parser.add_argument("--resource-file", type=Path, help="Optional JSON file with resource values")
parser.add_argument("--uptime-percent-30d", type=float)
parser.add_argument("--capacity-utilization", type=float)
parser.add_argument("--innovation", type=float)
parser.add_argument("--all-models-local", action="store_true")
parser.add_argument("--sovereign-stable-days", type=int)
parser.add_argument("--human-free-days", type=int)
parser.add_argument("--json", action="store_true")
parser.add_argument("--markdown", action="store_true", help="Render a markdown report instead of the terse CLI summary")
parser.add_argument("--output", type=Path, help="Optional file path for markdown output")
return parser.parse_args()
def _load_resources(args):
resources = dict(DEFAULT_RESOURCES)
if args.resource_file:
resources.update(json.loads(args.resource_file.read_text()))
overrides = {
"uptime_percent_30d": args.uptime_percent_30d,
"capacity_utilization": args.capacity_utilization,
"innovation": args.innovation,
"sovereign_stable_days": args.sovereign_stable_days,
"human_free_days": args.human_free_days,
}
for key, value in overrides.items():
if value is not None:
resources[key] = value
if args.all_models_local:
resources["all_models_local"] = True
return resources
def _load_issue_states(args, spec):
if args.issue_state_file:
raw = json.loads(args.issue_state_file.read_text())
return {int(k): v for k, v in raw.items()}
return load_issue_states(spec, token_file=args.token_file)
def _render_cli_summary(result: dict[str, Any]) -> str:
lines = [
"--- Fleet Progression Evaluator ---",
f"Epic #{result['epic_issue']}: {result['epic_title']}",
f"Current phase: {result['current_phase']['number']}{result['current_phase']['name']}",
f"Epic complete: {result['epic_complete']}",
]
if result["next_locked_phase"]:
lines.append(
f"Next locked phase: {result['next_locked_phase']['number']}{result['next_locked_phase']['name']}"
)
lines.append("")
for phase in result["phases"]:
lines.append(f"Phase {phase['number']} [{phase['status']}] {phase['name']}")
if phase["blocking_requirements"]:
for blocker in phase["blocking_requirements"]:
lines.append(
f" - blocked by {blocker['rule']}: actual={blocker['actual']} expected={blocker['expected']}"
)
return "\n".join(lines)
def main():
args = parse_args()
spec = load_spec(args.spec_file)
issue_states = _load_issue_states(args, spec)
resources = _load_resources(args)
result = evaluate_progression(spec, issue_states, resources, repo_root=DEFAULT_REPO_ROOT)
if args.json:
rendered = json.dumps(result, indent=2)
elif args.markdown or args.output:
rendered = render_markdown(result)
else:
rendered = _render_cli_summary(result)
if args.output:
args.output.parent.mkdir(parents=True, exist_ok=True)
args.output.write_text(rendered, encoding="utf-8")
print(f"Fleet progression report written to {args.output}")
else:
print(rendered)
if __name__ == "__main__":
main()

Some files were not shown because too many files have changed in this diff Show More