Compare commits
1 Commits
burn/667-1
...
fix/694
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
b334139fb5 |
21
ansible/inventory/group_vars/fleet.yml
Normal file
21
ansible/inventory/group_vars/fleet.yml
Normal file
@@ -0,0 +1,21 @@
|
||||
fleet_rotation_backup_root: /var/lib/timmy/secret-rotations
|
||||
fleet_secret_targets:
|
||||
ezra:
|
||||
env_file: /root/wizards/ezra/home/.env
|
||||
ssh_authorized_keys_file: /root/.ssh/authorized_keys
|
||||
services:
|
||||
- hermes-ezra.service
|
||||
- openclaw-ezra.service
|
||||
required_env_keys:
|
||||
- GITEA_TOKEN
|
||||
- TELEGRAM_BOT_TOKEN
|
||||
- PRIMARY_MODEL_API_KEY
|
||||
bezalel:
|
||||
env_file: /root/wizards/bezalel/home/.env
|
||||
ssh_authorized_keys_file: /root/.ssh/authorized_keys
|
||||
services:
|
||||
- hermes-bezalel.service
|
||||
required_env_keys:
|
||||
- GITEA_TOKEN
|
||||
- TELEGRAM_BOT_TOKEN
|
||||
- PRIMARY_MODEL_API_KEY
|
||||
79
ansible/inventory/group_vars/fleet_secrets.vault.yml
Normal file
79
ansible/inventory/group_vars/fleet_secrets.vault.yml
Normal file
@@ -0,0 +1,79 @@
|
||||
fleet_secret_bundle:
|
||||
ezra:
|
||||
env:
|
||||
GITEA_TOKEN: !vault |
|
||||
$ANSIBLE_VAULT;1.1;AES256
|
||||
38376433613738323463663336616263373734343839343866373561333334616233356531306361
|
||||
6334343162303937303834393664343033383765346666300a333236616231616461316436373430
|
||||
33316366656365663036663162616330616232653638376134373562356463653734613030333461
|
||||
3136633833656364640a646437626131316237646139663666313736666266613465323966646137
|
||||
33363735316239623130366266313466626262623137353331373430303930383931
|
||||
TELEGRAM_BOT_TOKEN: !vault |
|
||||
$ANSIBLE_VAULT;1.1;AES256
|
||||
35643034633034343630386637326166303264373838356635656330313762386339363232383363
|
||||
3136316263363738666133653965323530376231623633310a376138636662313366303435636465
|
||||
66303638376239623432613531633934313234663663366364373532346137356530613961363263
|
||||
6633393339356366380a393234393564353364373564363734626165386137343963303162356539
|
||||
33656137313463326534346138396365663536376561666132346534333234386266613562616135
|
||||
3764333036363165306165623039313239386362323030313032
|
||||
PRIMARY_MODEL_API_KEY: !vault |
|
||||
$ANSIBLE_VAULT;1.1;AES256
|
||||
61356337353033343634626430653031383161666130326135623134653736343732643364333762
|
||||
3532383230383337663632366235333230633430393238620a333962363730623735616137323833
|
||||
61343564346563313637303532626635373035396366636432366562666537613131653963663463
|
||||
6665613938313131630a343766383965393832386338333936653639343436666162613162356430
|
||||
31336264393536333963376632643135313164336637663564623336613032316561386566663538
|
||||
6330313233363564323462396561636165326562346333633664
|
||||
ssh_authorized_keys: !vault |
|
||||
$ANSIBLE_VAULT;1.1;AES256
|
||||
62373664326236626234643862666635393965656231366531633536626438396662663230343463
|
||||
3931666564356139386465346533353132396236393231640a656162633464653338613364626438
|
||||
39646232316637343662383631363533316432616161343734626235346431306532393337303362
|
||||
3964623239346166370a393330636134393535353730666165356131646332633937333062616536
|
||||
35376639346433383466346534343534373739643430313761633137636131313536383830656630
|
||||
34616335313836346435326665653732666238373232626335303336656462306434373432366366
|
||||
64323439366364663931386239303237633862633531666661313265613863376334323336333537
|
||||
31303434366237386362336535653561613963656137653330316431616466306262663237303366
|
||||
66353433666235613864346163393466383662313836626532663139623166346461313961363664
|
||||
31363136623830393439613038303465633138363933633364323035313332396366636463633134
|
||||
39653530386235363539313764303932643035373831326133396634303930346465663362643432
|
||||
37383236636262376165
|
||||
bezalel:
|
||||
env:
|
||||
GITEA_TOKEN: !vault |
|
||||
$ANSIBLE_VAULT;1.1;AES256
|
||||
64306432313532316331636139346633613930356232363238333037663038613038633937323266
|
||||
6661373032663265633662663532623736386433353737360a396531356230333761363836356436
|
||||
39653638343762633438333039366337346435663833613761313336666435373534363536376561
|
||||
6161633564326432350a623463633936373436636565643436336464343865613035633931376636
|
||||
65353666393830643536623764306236363462663130633835626337336531333932
|
||||
TELEGRAM_BOT_TOKEN: !vault |
|
||||
$ANSIBLE_VAULT;1.1;AES256
|
||||
37626132323238323938643034333634653038346239343062616638666163313266383365613530
|
||||
3838643864656265393830356632326630346237323133660a373361663265373366616636386233
|
||||
62306431646132363062633139653036643130333261366164393562633162366639636231313232
|
||||
6534303632653964350a343030333933623037656332626438323565626565616630623437386233
|
||||
65396233653434326563363738383035396235316233643934626332303435326562366261663435
|
||||
6333393861336535313637343037656135353339333935633762
|
||||
PRIMARY_MODEL_API_KEY: !vault |
|
||||
$ANSIBLE_VAULT;1.1;AES256
|
||||
31326537396565353334653537613938303566643561613365396665356139376433633564666364
|
||||
3266613539346234666165353633333539323537613535330a343734313438333566336638663466
|
||||
61353366303362333236383032363331323666386562383266613337393338356339323734633735
|
||||
6561666638376232320a386535373838633233373433366635393631396131336634303933326635
|
||||
30646232613466353666333034393462636331636430363335383761396561333630353639393633
|
||||
6363383263383734303534333437646663383233306333323336
|
||||
ssh_authorized_keys: !vault |
|
||||
$ANSIBLE_VAULT;1.1;AES256
|
||||
63643135646532323366613431616262653363636238376636666539393431623832343336383266
|
||||
3533666434356166366534336265343335663861313234650a393431383861346432396465363434
|
||||
33373737373130303537343061366134333138383735333538616637366561343337656332613237
|
||||
3736396561633734310a626637653634383134633137363630653966303765356665383832326663
|
||||
38613131353237623033656238373130633462363637646134373563656136623663366363343864
|
||||
37653563643030393531333766353665636163626637333336363664363930653437636338373564
|
||||
39313765393130383439653362663462666562376136396631626462653363303261626637333862
|
||||
31363664653535626236353330343834316661316533626433383230633236313762363235643737
|
||||
30313237303935303134656538343638633930333632653031383063363063353033353235323038
|
||||
36336361313661613465636335663964373636643139353932313663333231623466326332623062
|
||||
33646333626465373231653330323635333866303132633334393863306539643865656635376465
|
||||
65646434363538383035
|
||||
3
ansible/inventory/hosts.ini
Normal file
3
ansible/inventory/hosts.ini
Normal file
@@ -0,0 +1,3 @@
|
||||
[fleet]
|
||||
ezra ansible_host=143.198.27.163 ansible_user=root
|
||||
bezalel ansible_host=67.205.155.108 ansible_user=root
|
||||
185
ansible/playbooks/rotate_fleet_secrets.yml
Normal file
185
ansible/playbooks/rotate_fleet_secrets.yml
Normal file
@@ -0,0 +1,185 @@
|
||||
---
|
||||
- name: Rotate vaulted fleet secrets
|
||||
hosts: fleet
|
||||
gather_facts: false
|
||||
any_errors_fatal: true
|
||||
serial: 100%
|
||||
vars_files:
|
||||
- ../inventory/group_vars/fleet_secrets.vault.yml
|
||||
vars:
|
||||
rotation_id: "{{ lookup('pipe', 'date +%Y%m%d%H%M%S') }}"
|
||||
backup_root: "{{ fleet_rotation_backup_root }}/{{ rotation_id }}/{{ inventory_hostname }}"
|
||||
env_file_path: "{{ fleet_secret_targets[inventory_hostname].env_file }}"
|
||||
ssh_authorized_keys_path: "{{ fleet_secret_targets[inventory_hostname].ssh_authorized_keys_file }}"
|
||||
env_backup_path: "{{ backup_root }}/env.before"
|
||||
ssh_backup_path: "{{ backup_root }}/authorized_keys.before"
|
||||
staged_env_path: "{{ backup_root }}/env.candidate"
|
||||
staged_ssh_path: "{{ backup_root }}/authorized_keys.candidate"
|
||||
|
||||
tasks:
|
||||
- name: Validate target metadata and vaulted secret bundle
|
||||
ansible.builtin.assert:
|
||||
that:
|
||||
- fleet_secret_targets[inventory_hostname] is defined
|
||||
- fleet_secret_bundle[inventory_hostname] is defined
|
||||
- fleet_secret_targets[inventory_hostname].services | length > 0
|
||||
- fleet_secret_targets[inventory_hostname].required_env_keys | length > 0
|
||||
- fleet_secret_bundle[inventory_hostname].env is defined
|
||||
- fleet_secret_bundle[inventory_hostname].ssh_authorized_keys is defined
|
||||
- >-
|
||||
(fleet_secret_targets[inventory_hostname].required_env_keys
|
||||
| difference(fleet_secret_bundle[inventory_hostname].env.keys() | list)
|
||||
| length) == 0
|
||||
fail_msg: "rotation inventory incomplete for {{ inventory_hostname }}"
|
||||
|
||||
- name: Create backup directory for rotation bundle
|
||||
ansible.builtin.file:
|
||||
path: "{{ backup_root }}"
|
||||
state: directory
|
||||
mode: '0700'
|
||||
|
||||
- name: Check current env file
|
||||
ansible.builtin.stat:
|
||||
path: "{{ env_file_path }}"
|
||||
register: env_stat
|
||||
|
||||
- name: Check current authorized_keys file
|
||||
ansible.builtin.stat:
|
||||
path: "{{ ssh_authorized_keys_path }}"
|
||||
register: ssh_stat
|
||||
|
||||
- name: Read current env file
|
||||
ansible.builtin.slurp:
|
||||
src: "{{ env_file_path }}"
|
||||
register: env_current
|
||||
when: env_stat.stat.exists
|
||||
|
||||
- name: Read current authorized_keys file
|
||||
ansible.builtin.slurp:
|
||||
src: "{{ ssh_authorized_keys_path }}"
|
||||
register: ssh_current
|
||||
when: ssh_stat.stat.exists
|
||||
|
||||
- name: Save env rollback snapshot
|
||||
ansible.builtin.copy:
|
||||
content: "{{ env_current.content | b64decode }}"
|
||||
dest: "{{ env_backup_path }}"
|
||||
mode: '0600'
|
||||
when: env_stat.stat.exists
|
||||
|
||||
- name: Save authorized_keys rollback snapshot
|
||||
ansible.builtin.copy:
|
||||
content: "{{ ssh_current.content | b64decode }}"
|
||||
dest: "{{ ssh_backup_path }}"
|
||||
mode: '0600'
|
||||
when: ssh_stat.stat.exists
|
||||
|
||||
- name: Build staged env candidate
|
||||
ansible.builtin.copy:
|
||||
content: "{{ (env_current.content | b64decode) if env_stat.stat.exists else '' }}"
|
||||
dest: "{{ staged_env_path }}"
|
||||
mode: '0600'
|
||||
|
||||
- name: Stage rotated env secrets
|
||||
ansible.builtin.lineinfile:
|
||||
path: "{{ staged_env_path }}"
|
||||
regexp: "^{{ item.key }}="
|
||||
line: "{{ item.key }}={{ item.value }}"
|
||||
create: true
|
||||
loop: "{{ fleet_secret_bundle[inventory_hostname].env | dict2items }}"
|
||||
loop_control:
|
||||
label: "{{ item.key }}"
|
||||
no_log: true
|
||||
|
||||
- name: Ensure SSH directory exists
|
||||
ansible.builtin.file:
|
||||
path: "{{ ssh_authorized_keys_path | dirname }}"
|
||||
state: directory
|
||||
mode: '0700'
|
||||
|
||||
- name: Stage rotated authorized_keys bundle
|
||||
ansible.builtin.copy:
|
||||
content: "{{ fleet_secret_bundle[inventory_hostname].ssh_authorized_keys | trim ~ '\n' }}"
|
||||
dest: "{{ staged_ssh_path }}"
|
||||
mode: '0600'
|
||||
no_log: true
|
||||
|
||||
- name: Promote staged bundle, restart services, and verify health
|
||||
block:
|
||||
- name: Promote staged env file
|
||||
ansible.builtin.copy:
|
||||
src: "{{ staged_env_path }}"
|
||||
dest: "{{ env_file_path }}"
|
||||
remote_src: true
|
||||
mode: '0600'
|
||||
|
||||
- name: Promote staged authorized_keys
|
||||
ansible.builtin.copy:
|
||||
src: "{{ staged_ssh_path }}"
|
||||
dest: "{{ ssh_authorized_keys_path }}"
|
||||
remote_src: true
|
||||
mode: '0600'
|
||||
|
||||
- name: Restart dependent services
|
||||
ansible.builtin.systemd:
|
||||
name: "{{ item }}"
|
||||
state: restarted
|
||||
daemon_reload: true
|
||||
loop: "{{ fleet_secret_targets[inventory_hostname].services }}"
|
||||
loop_control:
|
||||
label: "{{ item }}"
|
||||
|
||||
- name: Verify service is active after restart
|
||||
ansible.builtin.command: "systemctl is-active {{ item }}"
|
||||
register: service_status
|
||||
changed_when: false
|
||||
failed_when: service_status.stdout.strip() != 'active'
|
||||
loop: "{{ fleet_secret_targets[inventory_hostname].services }}"
|
||||
loop_control:
|
||||
label: "{{ item }}"
|
||||
retries: 5
|
||||
delay: 2
|
||||
until: service_status.stdout.strip() == 'active'
|
||||
|
||||
rescue:
|
||||
- name: Restore env file from rollback snapshot
|
||||
ansible.builtin.copy:
|
||||
src: "{{ env_backup_path }}"
|
||||
dest: "{{ env_file_path }}"
|
||||
remote_src: true
|
||||
mode: '0600'
|
||||
when: env_stat.stat.exists
|
||||
|
||||
- name: Remove created env file when there was no prior version
|
||||
ansible.builtin.file:
|
||||
path: "{{ env_file_path }}"
|
||||
state: absent
|
||||
when: not env_stat.stat.exists
|
||||
|
||||
- name: Restore authorized_keys from rollback snapshot
|
||||
ansible.builtin.copy:
|
||||
src: "{{ ssh_backup_path }}"
|
||||
dest: "{{ ssh_authorized_keys_path }}"
|
||||
remote_src: true
|
||||
mode: '0600'
|
||||
when: ssh_stat.stat.exists
|
||||
|
||||
- name: Remove created authorized_keys when there was no prior version
|
||||
ansible.builtin.file:
|
||||
path: "{{ ssh_authorized_keys_path }}"
|
||||
state: absent
|
||||
when: not ssh_stat.stat.exists
|
||||
|
||||
- name: Restart services after rollback
|
||||
ansible.builtin.systemd:
|
||||
name: "{{ item }}"
|
||||
state: restarted
|
||||
daemon_reload: true
|
||||
loop: "{{ fleet_secret_targets[inventory_hostname].services }}"
|
||||
loop_control:
|
||||
label: "{{ item }}"
|
||||
ignore_errors: true
|
||||
|
||||
- name: Fail the rotation after rollback
|
||||
ansible.builtin.fail:
|
||||
msg: "Rotation failed for {{ inventory_hostname }}. Previous secrets restored from {{ backup_root }}."
|
||||
68
docs/FLEET_SECRET_ROTATION.md
Normal file
68
docs/FLEET_SECRET_ROTATION.md
Normal file
@@ -0,0 +1,68 @@
|
||||
# Fleet Secret Rotation
|
||||
|
||||
Issue: `timmy-home#694`
|
||||
|
||||
This runbook adds a single place to rotate fleet API keys, service tokens, and SSH authorized keys without hand-editing remote hosts.
|
||||
|
||||
## Files
|
||||
|
||||
- `ansible/inventory/hosts.ini` — fleet hosts (`ezra`, `bezalel`)
|
||||
- `ansible/inventory/group_vars/fleet.yml` — non-secret per-host targets (env file, services, authorized_keys path)
|
||||
- `ansible/inventory/group_vars/fleet_secrets.vault.yml` — vaulted `fleet_secret_bundle`
|
||||
- `ansible/playbooks/rotate_fleet_secrets.yml` — staged rotation + restart verification + rollback
|
||||
|
||||
## Secret inventory shape
|
||||
|
||||
`fleet_secret_bundle` is keyed by host. Each host carries the env secrets to rewrite plus the full `authorized_keys` payload to distribute.
|
||||
|
||||
```yaml
|
||||
fleet_secret_bundle:
|
||||
ezra:
|
||||
env:
|
||||
GITEA_TOKEN: !vault |
|
||||
...
|
||||
TELEGRAM_BOT_TOKEN: !vault |
|
||||
...
|
||||
PRIMARY_MODEL_API_KEY: !vault |
|
||||
...
|
||||
ssh_authorized_keys: !vault |
|
||||
...
|
||||
```
|
||||
|
||||
The committed vault file contains placeholder encrypted values only. Replace them with real rotated material before production use.
|
||||
|
||||
## Rotate a new bundle
|
||||
|
||||
From repo root:
|
||||
|
||||
```bash
|
||||
cd ansible
|
||||
ansible-vault edit inventory/group_vars/fleet_secrets.vault.yml
|
||||
ansible-playbook -i inventory/hosts.ini playbooks/rotate_fleet_secrets.yml --ask-vault-pass
|
||||
```
|
||||
|
||||
Or update one value at a time with `ansible-vault encrypt_string` and paste it into `fleet_secret_bundle`.
|
||||
|
||||
## What the playbook does
|
||||
|
||||
1. Validates that each host has a secret bundle and target metadata.
|
||||
2. Writes rollback snapshots under `/var/lib/timmy/secret-rotations/<rotation_id>/<host>/`.
|
||||
3. Stages a candidate `.env` file and candidate `authorized_keys` file before promotion.
|
||||
4. Promotes staged files into place.
|
||||
5. Restarts every declared dependent service.
|
||||
6. Verifies each service with `systemctl is-active`.
|
||||
7. If anything fails, restores the previous `.env` and `authorized_keys`, restarts services again, and aborts the run.
|
||||
|
||||
## Rollback semantics
|
||||
|
||||
Rollback is host-safe and automatic inside the playbook `rescue:` block.
|
||||
|
||||
- Existing `.env` and `authorized_keys` files are restored from backup when they existed before rotation.
|
||||
- Newly created files are removed if the host had no prior version.
|
||||
- Service restart is retried after rollback so the node returns to the last-known-good bundle.
|
||||
|
||||
## Operational notes
|
||||
|
||||
- Keep `required_env_keys` in `ansible/inventory/group_vars/fleet.yml` aligned with each house's real runtime contract.
|
||||
- `ssh_authorized_keys` distributes public keys only. Rotate corresponding private keys out-of-band, then publish the new authorized key list through the vault.
|
||||
- Use one vault edit per rotation window so API keys, bot tokens, and SSH access move together.
|
||||
@@ -9,6 +9,7 @@ Quick-reference index for common operational tasks across the Timmy Foundation i
|
||||
| Task | Location | Command/Procedure |
|
||||
|------|----------|-------------------|
|
||||
| Deploy fleet update | fleet-ops | `ansible-playbook playbooks/provision_and_deploy.yml --ask-vault-pass` |
|
||||
| Rotate fleet secrets | timmy-home | `cd ansible && ansible-playbook -i inventory/hosts.ini playbooks/rotate_fleet_secrets.yml --ask-vault-pass` |
|
||||
| Check fleet health | fleet-ops | `python3 scripts/fleet_readiness.py` |
|
||||
| Agent scorecard | fleet-ops | `python3 scripts/agent_scorecard.py` |
|
||||
| View fleet manifest | fleet-ops | `cat manifest.yaml` |
|
||||
|
||||
@@ -1,219 +0,0 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
Codebase Genome — Test Suite Generator
|
||||
|
||||
Scans a Python codebase, identifies uncovered functions/methods,
|
||||
and generates pytest test cases to fill coverage gaps.
|
||||
|
||||
Usage:
|
||||
python codebase-genome.py <target_dir> [--output tests/test_genome_generated.py]
|
||||
python codebase-genome.py <target_dir> --dry-run
|
||||
python codebase-genome.py <target_dir> --coverage
|
||||
"""
|
||||
|
||||
import ast
|
||||
import os
|
||||
import sys
|
||||
import argparse
|
||||
import subprocess
|
||||
import json
|
||||
from pathlib import Path
|
||||
from typing import List, Dict, Any, Optional, Set
|
||||
from dataclasses import dataclass, field
|
||||
|
||||
|
||||
@dataclass
|
||||
class FunctionInfo:
|
||||
name: str
|
||||
module: str
|
||||
file_path: str
|
||||
line_number: int
|
||||
is_method: bool = False
|
||||
class_name: Optional[str] = None
|
||||
args: List[str] = field(default_factory=list)
|
||||
has_return: bool = False
|
||||
raises: List[str] = field(default_factory=list)
|
||||
docstring: Optional[str] = None
|
||||
is_private: bool = False
|
||||
is_test: bool = False
|
||||
|
||||
|
||||
class CodebaseScanner:
|
||||
def __init__(self, target_dir: str):
|
||||
self.target_dir = Path(target_dir).resolve()
|
||||
self.functions: List[FunctionInfo] = []
|
||||
self.modules: Dict[str, List[FunctionInfo]] = {}
|
||||
|
||||
def scan(self) -> List[FunctionInfo]:
|
||||
for py_file in self.target_dir.rglob("*.py"):
|
||||
if self._should_skip(py_file):
|
||||
continue
|
||||
try:
|
||||
self._scan_file(py_file)
|
||||
except SyntaxError:
|
||||
print(f"Warning: Syntax error in {py_file}, skipping", file=sys.stderr)
|
||||
return self.functions
|
||||
|
||||
def _should_skip(self, path: Path) -> bool:
|
||||
skip_dirs = {"__pycache__", ".git", ".venv", "venv", "node_modules", ".tox"}
|
||||
if set(path.parts) & skip_dirs:
|
||||
return True
|
||||
if path.name.startswith("test_") or path.name.endswith("_test.py"):
|
||||
return True
|
||||
if path.name in ("conftest.py", "setup.py"):
|
||||
return True
|
||||
return False
|
||||
|
||||
def _scan_file(self, file_path: Path):
|
||||
content = file_path.read_text(encoding="utf-8", errors="replace")
|
||||
tree = ast.parse(content)
|
||||
module_name = self._get_module_name(file_path)
|
||||
|
||||
for node in ast.walk(tree):
|
||||
if isinstance(node, (ast.FunctionDef, ast.AsyncFunctionDef)):
|
||||
func = self._extract(node, module_name, file_path)
|
||||
if func and not func.is_test:
|
||||
self.functions.append(func)
|
||||
self.modules.setdefault(module_name, []).append(func)
|
||||
|
||||
def _get_module_name(self, file_path: Path) -> str:
|
||||
rel = file_path.relative_to(self.target_dir)
|
||||
parts = list(rel.parts)
|
||||
if parts[-1] == "__init__.py":
|
||||
parts = parts[:-1]
|
||||
else:
|
||||
parts[-1] = parts[-1].replace(".py", "")
|
||||
return ".".join(parts)
|
||||
|
||||
def _extract(self, node, module_name: str, file_path: Path) -> Optional[FunctionInfo]:
|
||||
if node.name.startswith("test_"):
|
||||
return None
|
||||
|
||||
args = [a.arg for a in node.args.args if a.arg not in ("self", "cls")]
|
||||
has_return = any(isinstance(n, ast.Return) and n.value for n in ast.walk(node))
|
||||
raises = []
|
||||
for n in ast.walk(node):
|
||||
if isinstance(n, ast.Raise) and n.exc and isinstance(n.exc, ast.Call):
|
||||
if isinstance(n.exc.func, ast.Name):
|
||||
raises.append(n.exc.func.id)
|
||||
|
||||
docstring = ast.get_docstring(node)
|
||||
is_method = False
|
||||
class_name = None
|
||||
for parent in ast.walk(tree := ast.parse(open(file_path).read())):
|
||||
for child in ast.iter_child_nodes(parent):
|
||||
if child is node and isinstance(parent, ast.ClassDef):
|
||||
is_method = True
|
||||
class_name = parent.name
|
||||
|
||||
return FunctionInfo(
|
||||
name=node.name, module=module_name, file_path=str(file_path),
|
||||
line_number=node.lineno, is_method=is_method, class_name=class_name,
|
||||
args=args, has_return=has_return, raises=raises, docstring=docstring,
|
||||
is_private=node.name.startswith("_") and not node.name.startswith("__"),
|
||||
)
|
||||
|
||||
|
||||
class TestGenerator:
|
||||
HEADER = '''# AUTO-GENERATED by codebase-genome.py — review before committing
|
||||
|
||||
import pytest
|
||||
from unittest.mock import patch, MagicMock
|
||||
import sys
|
||||
from pathlib import Path
|
||||
|
||||
sys.path.insert(0, str(Path(__file__).resolve().parents[1]))
|
||||
|
||||
'''
|
||||
|
||||
def generate(self, functions: List[FunctionInfo]) -> str:
|
||||
parts = [self.HEADER]
|
||||
modules: Dict[str, List[FunctionInfo]] = {}
|
||||
for f in functions:
|
||||
modules.setdefault(f.module, []).append(f)
|
||||
|
||||
for mod, funcs in sorted(modules.items()):
|
||||
parts.append(f"# ═══ {mod} ═══\n")
|
||||
imp = mod.replace("-", "_")
|
||||
parts.append(f"try:\n from {imp} import *\nexcept ImportError:\n pytest.skip('{imp} not importable', allow_module_level=True)\n")
|
||||
|
||||
for func in funcs:
|
||||
test = self._gen_test(func)
|
||||
if test:
|
||||
parts.append(test + "\n")
|
||||
|
||||
return "\n".join(parts)
|
||||
|
||||
def _gen_test(self, func: FunctionInfo) -> Optional[str]:
|
||||
name = f"test_{func.module.replace('.', '_')}_{func.name}"
|
||||
lines = [f"def {name}():", f' """Auto-generated for {func.module}.{func.name}."""']
|
||||
|
||||
if not func.args:
|
||||
lines += [
|
||||
" try:",
|
||||
f" r = {func.name}()",
|
||||
" assert r is not None or r is None",
|
||||
" except Exception:",
|
||||
" pass",
|
||||
]
|
||||
else:
|
||||
lines += [
|
||||
" try:",
|
||||
f" {func.name}({', '.join(a + '=None' for a in func.args)})",
|
||||
" except (TypeError, ValueError, AttributeError):",
|
||||
" pass",
|
||||
]
|
||||
if any(a in ("text", "content", "message", "query", "path") for a in func.args):
|
||||
lines += [
|
||||
" try:",
|
||||
f" {func.name}({', '.join(a + '=\"\"' if a in ('text','content','message','query','path') else a + '=None' for a in func.args)})",
|
||||
" except (TypeError, ValueError):",
|
||||
" pass",
|
||||
]
|
||||
|
||||
if func.raises:
|
||||
lines.append(f" # May raise: {', '.join(func.raises[:2])}")
|
||||
lines.append(f" # with pytest.raises(({', '.join(func.raises[:2])})):")
|
||||
lines.append(f" # {func.name}()")
|
||||
|
||||
return "\n".join(lines)
|
||||
|
||||
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(description="Codebase Genome — Test Generator")
|
||||
parser.add_argument("target_dir")
|
||||
parser.add_argument("--output", "-o", default="tests/test_genome_generated.py")
|
||||
parser.add_argument("--dry-run", action="store_true")
|
||||
parser.add_argument("--max-tests", type=int, default=100)
|
||||
args = parser.parse_args()
|
||||
|
||||
target = Path(args.target_dir).resolve()
|
||||
if not target.is_dir():
|
||||
print(f"Error: {target} not a directory", file=sys.stderr)
|
||||
return 1
|
||||
|
||||
print(f"Scanning {target}...")
|
||||
scanner = CodebaseScanner(str(target))
|
||||
functions = scanner.scan()
|
||||
print(f"Found {len(functions)} functions in {len(scanner.modules)} modules")
|
||||
|
||||
if len(functions) > args.max_tests:
|
||||
print(f"Limiting to {args.max_tests}")
|
||||
functions = functions[:args.max_tests]
|
||||
|
||||
gen = TestGenerator()
|
||||
code = gen.generate(functions)
|
||||
|
||||
if args.dry_run:
|
||||
print(code)
|
||||
return 0
|
||||
|
||||
out = target / args.output
|
||||
out.parent.mkdir(parents=True, exist_ok=True)
|
||||
out.write_text(code)
|
||||
print(f"Generated {len(functions)} tests → {out}")
|
||||
return 0
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
sys.exit(main())
|
||||
87
tests/test_fleet_secret_rotation.py
Normal file
87
tests/test_fleet_secret_rotation.py
Normal file
@@ -0,0 +1,87 @@
|
||||
#!/usr/bin/env python3
|
||||
"""Regression coverage for timmy-home #694 fleet secret rotation assets."""
|
||||
|
||||
from pathlib import Path
|
||||
import unittest
|
||||
|
||||
import yaml
|
||||
|
||||
|
||||
ROOT = Path(__file__).resolve().parents[1]
|
||||
ANSIBLE_DIR = ROOT / "ansible"
|
||||
HOSTS_FILE = ANSIBLE_DIR / "inventory" / "hosts.ini"
|
||||
TARGETS_FILE = ANSIBLE_DIR / "inventory" / "group_vars" / "fleet.yml"
|
||||
SECRETS_FILE = ANSIBLE_DIR / "inventory" / "group_vars" / "fleet_secrets.vault.yml"
|
||||
PLAYBOOK_FILE = ANSIBLE_DIR / "playbooks" / "rotate_fleet_secrets.yml"
|
||||
DOC_FILE = ROOT / "docs" / "FLEET_SECRET_ROTATION.md"
|
||||
|
||||
|
||||
class TestFleetSecretRotation(unittest.TestCase):
|
||||
def test_inventory_declares_each_host_target(self):
|
||||
self.assertTrue(HOSTS_FILE.exists(), "missing ansible inventory hosts file")
|
||||
self.assertTrue(TARGETS_FILE.exists(), "missing fleet target metadata")
|
||||
|
||||
hosts_text = HOSTS_FILE.read_text(encoding="utf-8")
|
||||
self.assertIn("[fleet]", hosts_text)
|
||||
self.assertIn("ezra", hosts_text)
|
||||
self.assertIn("bezalel", hosts_text)
|
||||
|
||||
targets = yaml.safe_load(TARGETS_FILE.read_text(encoding="utf-8"))
|
||||
self.assertIn("fleet_secret_targets", targets)
|
||||
|
||||
expected_env_files = {
|
||||
"ezra": "/root/wizards/ezra/home/.env",
|
||||
"bezalel": "/root/wizards/bezalel/home/.env",
|
||||
}
|
||||
for host, env_file in expected_env_files.items():
|
||||
self.assertIn(host, targets["fleet_secret_targets"])
|
||||
target = targets["fleet_secret_targets"][host]
|
||||
self.assertEqual(target["env_file"], env_file)
|
||||
self.assertEqual(target["ssh_authorized_keys_file"], "/root/.ssh/authorized_keys")
|
||||
self.assertGreaterEqual(len(target["services"]), 1)
|
||||
self.assertGreaterEqual(len(target["required_env_keys"]), 3)
|
||||
|
||||
def test_vault_file_contains_encrypted_secret_bundle_for_each_host(self):
|
||||
self.assertTrue(SECRETS_FILE.exists(), "missing vaulted secrets inventory")
|
||||
text = SECRETS_FILE.read_text(encoding="utf-8")
|
||||
self.assertIn("fleet_secret_bundle:", text)
|
||||
self.assertIn("$ANSIBLE_VAULT;1.1;AES256", text)
|
||||
for host in ("ezra", "bezalel"):
|
||||
self.assertIn(f" {host}:", text)
|
||||
self.assertGreaterEqual(text.count("!vault |"), 4)
|
||||
|
||||
def test_playbook_has_staging_verification_and_rollback(self):
|
||||
self.assertTrue(PLAYBOOK_FILE.exists(), "missing rotation playbook")
|
||||
text = PLAYBOOK_FILE.read_text(encoding="utf-8")
|
||||
for snippet in (
|
||||
"any_errors_fatal: true",
|
||||
"vars_files:",
|
||||
"fleet_secrets.vault.yml",
|
||||
"backup_root",
|
||||
"env_backup_path",
|
||||
"ssh_backup_path",
|
||||
"lineinfile:",
|
||||
"copy:",
|
||||
"systemd:",
|
||||
"state: restarted",
|
||||
"systemctl is-active",
|
||||
"block:",
|
||||
"rescue:",
|
||||
):
|
||||
self.assertIn(snippet, text)
|
||||
|
||||
def test_docs_explain_rotation_command_and_rollback(self):
|
||||
self.assertTrue(DOC_FILE.exists(), "missing fleet secret rotation docs")
|
||||
text = DOC_FILE.read_text(encoding="utf-8")
|
||||
for snippet in (
|
||||
"ansible-playbook",
|
||||
"--ask-vault-pass",
|
||||
"rollback",
|
||||
"authorized_keys",
|
||||
"fleet_secret_bundle",
|
||||
):
|
||||
self.assertIn(snippet, text)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
unittest.main(verbosity=2)
|
||||
@@ -1,319 +0,0 @@
|
||||
# GENOME.md — the-nexus
|
||||
|
||||
**Generated:** 2026-04-14
|
||||
**Repo:** Timmy_Foundation/the-nexus
|
||||
**Analysis:** Codebase Genome #672
|
||||
|
||||
---
|
||||
|
||||
## Project Overview
|
||||
|
||||
The Nexus is Timmy's canonical 3D home-world — a browser-based Three.js application that serves as:
|
||||
1. **Local-first training ground** for Timmy (the sovereign AI)
|
||||
2. **Wizardly visualization surface** for the fleet system
|
||||
3. **Portal architecture** connecting to other worlds and services
|
||||
|
||||
The app is a real-time 3D environment with spatial memory, GOFAI reasoning, agent presence, and portal-based navigation.
|
||||
|
||||
---
|
||||
|
||||
## Architecture
|
||||
|
||||
```mermaid
|
||||
graph TB
|
||||
subgraph Browser["BROWSER LAYER"]
|
||||
HTML[index.html]
|
||||
APP[app.js - 4082 lines]
|
||||
CSS[style.css]
|
||||
Worker[gofai_worker.js]
|
||||
end
|
||||
|
||||
subgraph ThreeJS["THREE.JS RENDERING"]
|
||||
Scene[Scene Management]
|
||||
Camera[Camera System]
|
||||
Renderer[WebGL Renderer]
|
||||
Post[Post-processing<br/>Bloom, SMAA]
|
||||
Physics[Physics/Player]
|
||||
end
|
||||
|
||||
subgraph Nexus["NEXUS COMPONENTS"]
|
||||
SM[SpatialMemory]
|
||||
SA[SpatialAudio]
|
||||
MB[MemoryBirth]
|
||||
MO[MemoryOptimizer]
|
||||
MI[MemoryInspect]
|
||||
MP[MemoryPulse]
|
||||
RT[ReasoningTrace]
|
||||
RV[ResonanceVisualizer]
|
||||
end
|
||||
|
||||
subgraph GOFAI["GOFAI REASONING"]
|
||||
Worker2[Web Worker]
|
||||
Rules[Rule Engine]
|
||||
Facts[Fact Store]
|
||||
Inference[Inference Loop]
|
||||
end
|
||||
|
||||
subgraph Backend["BACKEND SERVICES"]
|
||||
Server[server.py<br/>WebSocket Bridge]
|
||||
L402[L402 Cost API]
|
||||
Portal[Portal Registry]
|
||||
end
|
||||
|
||||
subgraph Data["DATA/PERSISTENCE"]
|
||||
Local[localStorage]
|
||||
IDB[IndexedDB]
|
||||
JSON[portals.json]
|
||||
Vision[vision.json]
|
||||
end
|
||||
|
||||
HTML --> APP
|
||||
APP --> ThreeJS
|
||||
APP --> Nexus
|
||||
APP --> GOFAI
|
||||
APP --> Backend
|
||||
APP --> Data
|
||||
|
||||
Worker2 --> APP
|
||||
Server --> APP
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Entry Points
|
||||
|
||||
### Primary Entry
|
||||
- **`index.html`** — Main HTML shell, loads app.js
|
||||
- **`app.js`** — Main application (4082 lines), Three.js scene setup
|
||||
|
||||
### Secondary Entry Points
|
||||
- **`boot.js`** — Bootstrap sequence
|
||||
- **`bootstrap.mjs`** — ES module bootstrap
|
||||
- **`server.py`** — WebSocket bridge server
|
||||
|
||||
### Configuration Entry Points
|
||||
- **`portals.json`** — Portal definitions and destinations
|
||||
- **`vision.json`** — Vision/agent configuration
|
||||
- **`config/fleet_agents.json`** — Fleet agent definitions
|
||||
|
||||
---
|
||||
|
||||
## Data Flow
|
||||
|
||||
```
|
||||
User Input
|
||||
↓
|
||||
app.js (Event Loop)
|
||||
↓
|
||||
┌─────────────────────────────────────┐
|
||||
│ Three.js Scene │
|
||||
│ - Player movement │
|
||||
│ - Camera controls │
|
||||
│ - Physics simulation │
|
||||
│ - Portal detection │
|
||||
└─────────────────────────────────────┘
|
||||
↓
|
||||
┌─────────────────────────────────────┐
|
||||
│ Nexus Components │
|
||||
│ - SpatialMemory (room/context) │
|
||||
│ - MemoryBirth (new memories) │
|
||||
│ - MemoryPulse (heartbeat) │
|
||||
│ - ReasoningTrace (GOFAI output) │
|
||||
└─────────────────────────────────────┘
|
||||
↓
|
||||
┌─────────────────────────────────────┐
|
||||
│ GOFAI Worker (off-thread) │
|
||||
│ - Rule evaluation │
|
||||
│ - Fact inference │
|
||||
│ - Decision making │
|
||||
└─────────────────────────────────────┘
|
||||
↓
|
||||
┌─────────────────────────────────────┐
|
||||
│ Backend Services │
|
||||
│ - WebSocket (server.py) │
|
||||
│ - L402 cost API │
|
||||
│ - Portal registry │
|
||||
└─────────────────────────────────────┘
|
||||
↓
|
||||
Persistence (localStorage/IndexedDB)
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Key Abstractions
|
||||
|
||||
### 1. Nexus Object (`NEXUS`)
|
||||
Central configuration and state object containing:
|
||||
- Color palette
|
||||
- Room definitions
|
||||
- Portal configurations
|
||||
- Agent settings
|
||||
|
||||
### 2. SpatialMemory
|
||||
Manages room-based context for the AI agent:
|
||||
- Room transitions trigger context switches
|
||||
- Facts are stored per-room
|
||||
- NPCs have location awareness
|
||||
|
||||
### 3. Portal System
|
||||
Connects the 3D world to external services:
|
||||
- Portals defined in `portals.json`
|
||||
- Each portal links to a service/endpoint
|
||||
- Visual indicators in 3D space
|
||||
|
||||
### 4. GOFAI Worker
|
||||
Off-thread reasoning engine:
|
||||
- Rule-based inference
|
||||
- Fact store with persistence
|
||||
- Decision making for agent behavior
|
||||
|
||||
### 5. Memory Components
|
||||
- **MemoryBirth**: Creates new memories from interactions
|
||||
- **MemoryOptimizer**: Compresses and deduplicates memories
|
||||
- **MemoryPulse**: Heartbeat system for memory health
|
||||
- **MemoryInspect**: Debug/inspection interface
|
||||
|
||||
---
|
||||
|
||||
## API Surface
|
||||
|
||||
### Internal APIs (JavaScript)
|
||||
|
||||
| Module | Export | Purpose |
|
||||
|--------|--------|---------|
|
||||
| `app.js` | `NEXUS` | Main config/state object |
|
||||
| `SpatialMemory` | class | Room-based context management |
|
||||
| `SpatialAudio` | class | 3D positional audio |
|
||||
| `MemoryBirth` | class | Memory creation |
|
||||
| `MemoryOptimizer` | class | Memory compression |
|
||||
| `ReasoningTrace` | class | GOFAI reasoning visualization |
|
||||
|
||||
### External APIs (HTTP/WebSocket)
|
||||
|
||||
| Endpoint | Protocol | Purpose |
|
||||
|----------|----------|---------|
|
||||
| `ws://localhost:PORT` | WebSocket | Real-time bridge to backend |
|
||||
| `http://localhost:8080/api/cost-estimate` | HTTP | L402 cost estimation |
|
||||
| Portal endpoints | Various | External service connections |
|
||||
|
||||
---
|
||||
|
||||
## Dependencies
|
||||
|
||||
### Runtime Dependencies
|
||||
- **Three.js** — 3D rendering engine
|
||||
- **Three.js Addons** — Post-processing (Bloom, SMAA)
|
||||
|
||||
### Build Dependencies
|
||||
- **ES Modules** — Native browser modules
|
||||
- **No bundler** — Direct script loading
|
||||
|
||||
### Backend Dependencies
|
||||
- **Python 3.x** — server.py
|
||||
- **WebSocket** — Real-time communication
|
||||
|
||||
---
|
||||
|
||||
## Test Coverage
|
||||
|
||||
### Existing Tests
|
||||
- `tests/boot.test.js` — Bootstrap sequence tests
|
||||
|
||||
### Test Gaps
|
||||
1. **Three.js scene initialization** — No tests
|
||||
2. **Portal system** — No tests
|
||||
3. **Memory components** — No tests
|
||||
4. **GOFAI worker** — No tests
|
||||
5. **WebSocket communication** — No tests
|
||||
6. **Spatial memory transitions** — No tests
|
||||
7. **Physics/player movement** — No tests
|
||||
|
||||
### Recommended Test Priorities
|
||||
1. Portal detection and activation
|
||||
2. Spatial memory room transitions
|
||||
3. GOFAI worker message passing
|
||||
4. WebSocket connection handling
|
||||
5. Memory persistence (localStorage/IndexedDB)
|
||||
|
||||
---
|
||||
|
||||
## Security Considerations
|
||||
|
||||
### Current Risks
|
||||
1. **WebSocket without auth** — server.py has no authentication
|
||||
2. **localStorage sensitive data** — Memories stored unencrypted
|
||||
3. **CORS open** — No origin restrictions on WebSocket
|
||||
4. **L402 endpoint** — Cost API may expose internal state
|
||||
|
||||
### Mitigations
|
||||
1. Add WebSocket authentication
|
||||
2. Encrypt sensitive memories
|
||||
3. Restrict CORS origins
|
||||
4. Rate limit L402 endpoint
|
||||
|
||||
---
|
||||
|
||||
## File Structure
|
||||
|
||||
```
|
||||
the-nexus/
|
||||
├── app.js # Main app (4082 lines)
|
||||
├── index.html # HTML shell
|
||||
├── style.css # Styles
|
||||
├── server.py # WebSocket bridge
|
||||
├── boot.js # Bootstrap
|
||||
├── bootstrap.mjs # ES module bootstrap
|
||||
├── gofai_worker.js # GOFAI web worker
|
||||
├── portals.json # Portal definitions
|
||||
├── vision.json # Vision config
|
||||
├── nexus/ # Nexus components
|
||||
│ └── components/
|
||||
│ ├── spatial-memory.js
|
||||
│ ├── spatial-audio.js
|
||||
│ ├── memory-birth.js
|
||||
│ ├── memory-optimizer.js
|
||||
│ ├── memory-inspect.js
|
||||
│ ├── memory-pulse.js
|
||||
│ ├── reasoning-trace.js
|
||||
│ └── resonance-visualizer.js
|
||||
├── config/ # Configuration
|
||||
├── docs/ # Documentation
|
||||
├── tests/ # Tests
|
||||
├── agent/ # Agent components
|
||||
├── bin/ # Scripts
|
||||
└── assets/ # Static assets
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## Technical Debt
|
||||
|
||||
1. **Large app.js** (4082 lines) — Should be split into modules
|
||||
2. **No TypeScript** — Pure JavaScript, no type safety
|
||||
3. **Manual DOM manipulation** — Could use a framework
|
||||
4. **No build system** — Direct ES modules, no optimization
|
||||
5. **Limited error handling** — Minimal try/catch coverage
|
||||
|
||||
---
|
||||
|
||||
## Migration Notes
|
||||
|
||||
From CLAUDE.md:
|
||||
- Current `main` does NOT ship the old root frontend files
|
||||
- A clean checkout serves a directory listing
|
||||
- The live browser shell exists in legacy form at `/Users/apayne/the-matrix`
|
||||
- Migration priorities: #684 (docs), #685 (legacy audit), #686 (smoke tests), #687 (restore shell)
|
||||
|
||||
---
|
||||
|
||||
## Next Steps
|
||||
|
||||
1. **Restore browser shell** — Bring frontend back to main
|
||||
2. **Add tests** — Cover critical paths (portals, memory, GOFAI)
|
||||
3. **Split app.js** — Modularize the 4082-line file
|
||||
4. **Add authentication** — Secure WebSocket and APIs
|
||||
5. **TypeScript migration** — Add type safety
|
||||
|
||||
---
|
||||
|
||||
*Generated by Codebase Genome pipeline — Issue #672*
|
||||
Reference in New Issue
Block a user