Compare commits

..

2 Commits

Author SHA1 Message Date
Alexander Whitestone
9339b1a2d4 docs: ground unified fleet sovereignty directive
Some checks failed
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 12s
Agent PR Gate / gate (pull_request) Failing after 31s
Smoke Test / smoke (pull_request) Failing after 16s
Agent PR Gate / report (pull_request) Successful in 9s
Refs #524
2026-04-22 10:29:16 -04:00
Alexander Whitestone
a6ec2055cc wip: add issue 524 directive grounding script 2026-04-22 10:27:30 -04:00
7 changed files with 617 additions and 257 deletions

View File

@@ -0,0 +1,107 @@
# [DIRECTIVE] Unified Fleet Sovereignty & Comms Migration
Grounding report for `timmy-home #524`.
Issue #524 is a multi-lane directive, not a one-commit feature. This report grounds the directive in repo evidence, highlights stale cross-links, and names the missing operator bundles that still need real execution.
This remains a `Refs #524` artifact. The directive spans multiple repos and operator actions, so this report makes the current repo-side state executable without pretending the whole migration is complete.
## Directive Snapshot
- Repo-grounded workstreams: 0
- Partial workstreams: 4
- Missing workstreams: 1
- Drifted references: 4
## Reference Drift
- #813 is cited for Nostr Migration Leadership, but its current title is 'docs: refresh the-playground genome analysis (#671)'.
- #819 is cited for Nostr Migration Leadership, but its current title is 'docs: verify #648 already implemented (closes #818)'.
- #139 is cited for v0.7.0 Feature Audit, but its current title is '🐣 Allegro-Primus is born'.
- #103 is cited for Morrowind Local-First Benchmark, but its current title is 'Build comprehensive caching layer — cache everywhere'.
## Workstream Matrix
### 1. Nostr Migration Leadership — PARTIAL
- Requirement: Replace Telegram with relay-based sovereign comms, verify wizard keypairs, and prove the NIP-29 group path is stable.
- Referenced issues:
- #813 (closed) — docs: refresh the-playground genome analysis (#671) [DRIFT]
- #819 (open) — docs: verify #648 already implemented (closes #818) [DRIFT]
- Repo evidence present:
- `infrastructure/timmy-bridge/client/timmy_client.py` — Nostr event client scaffold already exists
- `infrastructure/timmy-bridge/monitor/timmy_monitor.py` — Nostr relay monitor already exists
- `specs/wizard-telegram-bot-cutover.md` — Telegram cutover planning exists, so the migration lane is real
- Missing operator deliverables:
- wizard keypair inventory and ownership matrix
- NIP-29 relay group verification report
- operator runbook for cutting traffic off Telegram
- Why this lane remains open: The repo has Nostr-adjacent scaffolding, but the directive still lacks a verified migration packet and the cited issue links drift away from the stated Nostr scope.
### 2. Lexicon Enforcement — PARTIAL
- Requirement: Enforce the Fleet Lexicon in PR review and issue triage so the team uses one shared language.
- Referenced issues:
- #388 (closed) — [KT] Fleet Lexicon & Techniques — Shared Vocabulary, Patterns, and Standards for All Agents [aligned]
- Repo evidence present:
- `docs/WIZARD_APPRENTICESHIP_CHARTER.md` — The repo already uses wizard-language canon in docs
- `specs/timmy-ezra-bezalel-canon-sheet.md` — Canonical agent naming already exists
- `docs/OPERATIONS_DASHBOARD.md` — Operational roles are already described in repo language
- Missing operator deliverables:
- machine-checkable lexicon policy for review/triage
- terminology lint or reviewer checklist tied to the lexicon
- Why this lane remains open: The naming canon exists, but there is still no executable enforcement bundle that would catch drift during future reviews and triage passes.
### 3. v0.7.0 Feature Audit — PARTIAL
- Requirement: Audit Hermes features that can reduce cloud dependency and turn the findings into a sovereignty implementation plan.
- Referenced issues:
- #139 (open) — 🐣 Allegro-Primus is born [DRIFT]
- Repo evidence present:
- `scripts/sovereignty_audit.py` — Cloud-vs-local audit machinery already exists
- `reports/evaluations/2026-04-15-phase-4-sovereignty-audit.md` — Recent sovereignty audit report is committed
- `timmy-local/README.md` — Local-first status is already documented for operators
- Missing operator deliverables:
- Hermes v0.7.0 feature inventory linked to cloud-reduction leverage
- Sovereignty Implementation Plan derived from that feature audit
- Why this lane remains open: The repo has sovereignty-audit infrastructure, but it does not yet contain the requested v0.7.0 feature inventory or the plan that turns those findings into rollout steps.
### 4. Morrowind Local-First Benchmark — PARTIAL
- Requirement: Compare cloud and local Morrowind agents, prove local parity where possible, and document the reasoning gap when it fails.
- Referenced issues:
- #103 (open) — Build comprehensive caching layer — cache everywhere [DRIFT]
- Repo evidence present:
- `morrowind/local_brain.py` — Local Morrowind control loop already exists
- `morrowind/mcp_server.py` — Morrowind MCP control surface is already wired
- `morrowind/pilot.py` — Trajectory logging for evaluation already exists
- Missing operator deliverables:
- cloud-vs-local benchmark report for the combat loop
- reasoning-gap writeup tied to a proposed LoRA/fine-tune path
- Why this lane remains open: The repo has a local Morrowind stack, but it does not yet contain the requested benchmark artifact; the cited issue number also points at an unrelated caching task.
### 5. Infrastructure Hardening / Syntax Guard — MISSING
- Requirement: Verify Syntax Guard pre-receive protection across Gitea repos so syntax failures stop earlier.
- Referenced issues: none listed in the directive body
- Repo evidence present: none
- Missing operator deliverables:
- repo inventory of Gitea targets that should carry Syntax Guard
- deployment verifier for hook presence across those repos
- operator report proving installation state instead of assuming it
- Why this lane remains open: No repo-managed syntax-guard verifier is present yet, so this directive still depends on manual trust rather than auditable proof.
## Highest-Leverage Next Actions
- Nostr Migration Leadership: wizard keypair inventory and ownership matrix
- Lexicon Enforcement: machine-checkable lexicon policy for review/triage
- v0.7.0 Feature Audit: Hermes v0.7.0 feature inventory linked to cloud-reduction leverage
- Morrowind Local-First Benchmark: cloud-vs-local benchmark report for the combat loop
- Infrastructure Hardening / Syntax Guard: repo inventory of Gitea targets that should carry Syntax Guard
## Why #524 Remains Open
- The directive bundles five separate workstreams with different evidence surfaces.
- Multiple cited issue numbers have drifted away from the work they are supposed to anchor.
- Repo scaffolding exists for Nostr, sovereignty audits, and Morrowind, but the operator-facing bundles are still missing.
- Syntax Guard verification is still undocumented and unproven inside this repo.

View File

@@ -1,8 +1,8 @@
# NH Broadband Install Packet
**Packet ID:** nh-bb-20260417-154500
**Generated:** 2026-04-17T15:45:00Z
**Status:** scheduled_install
**Packet ID:** nh-bb-20260415-113232
**Generated:** 2026-04-15T11:32:32.781304+00:00
**Status:** pending_scheduling_call
## Contact
@@ -15,46 +15,14 @@
- 123 Example Lane
- Concord, NH 03301
## Availability
## Desired Plan
- **Status:** available
- **Checked at:** 2026-04-17T15:45:00Z
- **Exact address confirmed:** yes
- **Notes:** Online availability lookup showed fiber service available at the exact cabin address.
## Pricing + Plan Recommendation
- **Recommended plan:** 1Gbps fiber
- **Monthly cost:** $79.95
- **Install fee:** $99.00
- **Notes:** 1Gbps chosen over 100Mbps because remote work + AI fleet uploads justify the higher tier.
## Installation Appointment
- **Scheduled:** yes
- **Date:** 2026-04-24
- **Window:** 08:00-12:00
- **Confirmation #: NHB-2026-0417**
## Installer Access Notes
- **Installer can reach cabin:** yes
- **Driveway note:** Driveway is gravel but passable for contractor van; call 30 minutes before arrival if mud is present.
- **Site contact:** 603-555-0142
## Payment
- **Method:** credit_card
- **First month due:** $79.95
- **Install fee due:** $99.00
- **Notes:** Card on file approved for first month plus install fee.
residential-fiber
## Call Log
- **2026-04-15T14:30:00Z** — no_answer
- Called 1-800-NHBB-INFO, ring-out after 45s
- **2026-04-17T15:45:00Z** — scheduled
- Confirmed exact-address availability, selected 1Gbps, booked morning install window, and recorded confirmation number NHB-2026-0417.
## Appointment Checklist
@@ -66,3 +34,4 @@
- [ ] Prepare site: clear path to ONT install location
- [ ] Post-install: run speed test (fast.com / speedtest.net)
- [ ] Log final speeds and appointment outcome

View File

@@ -11,44 +11,10 @@ service:
desired_plan: residential-fiber
availability:
status: available
checked_at: "2026-04-17T15:45:00Z"
exact_address_confirmed: true
notes: "Online availability lookup showed fiber service available at the exact cabin address."
pricing:
recommended_plan: 1Gbps fiber
monthly_cost_usd: 79.95
install_fee_usd: 99.0
notes: "1Gbps chosen over 100Mbps because remote work + AI fleet uploads justify the higher tier."
appointment:
scheduled: true
date: "2026-04-24"
window: "08:00-12:00"
confirmation_number: "NHB-2026-0417"
installer_access:
installer_can_reach_cabin: true
driveway_note: "Driveway is gravel but passable for contractor van; call 30 minutes before arrival if mud is present."
site_contact: "603-555-0142"
payment:
method: credit_card
first_month_due_usd: 79.95
install_fee_due_usd: 99.0
notes: "Card on file approved for first month plus install fee."
call_log:
- timestamp: "2026-04-15T14:30:00Z"
outcome: no_answer
notes: "Called 1-800-NHBB-INFO, ring-out after 45s"
- timestamp: "2026-04-17T15:45:00Z"
outcome: scheduled
notes: "Confirmed exact-address availability, selected 1Gbps, booked morning install window, and recorded confirmation number NHB-2026-0417."
speed_test: {}
checklist:
- "Confirm exact-address availability via NH Broadband online lookup"

View File

@@ -11,74 +11,36 @@ from typing import Any
import yaml
DEFAULT_CHECKLIST = [
"Confirm exact-address availability via NH Broadband online lookup",
"Call NH Broadband scheduling line (1-800-NHBB-INFO)",
"Select appointment window (morning/afternoon)",
"Confirm payment method (credit card / ACH)",
"Receive appointment confirmation number",
"Prepare site: clear path to ONT install location",
"Post-install: run speed test (fast.com / speedtest.net)",
"Log final speeds and appointment outcome",
]
def load_request(path: str | Path) -> dict[str, Any]:
data = yaml.safe_load(Path(path).read_text()) or {}
data.setdefault("contact", {})
data.setdefault("service", {})
data.setdefault("call_log", [])
data.setdefault("checklist", list(DEFAULT_CHECKLIST))
data.setdefault("availability", {})
data.setdefault("pricing", {})
data.setdefault("appointment", {})
data.setdefault("installer_access", {})
data.setdefault("payment", {})
data.setdefault("speed_test", {})
data.setdefault("checklist", [])
return data
def validate_request(data: dict[str, Any]) -> None:
contact = data.get("contact", {})
for field in ("name", "phone"):
if not str(contact.get(field, "")).strip():
if not contact.get(field, "").strip():
raise ValueError(f"contact.{field} is required")
service = data.get("service", {})
for field in ("address", "city", "state"):
if not str(service.get(field, "")).strip():
if not service.get(field, "").strip():
raise ValueError(f"service.{field} is required")
if not data.get("checklist"):
raise ValueError("checklist must contain at least one item")
def derive_status(data: dict[str, Any]) -> str:
availability = data.get("availability", {})
appointment = data.get("appointment", {})
speed_test = data.get("speed_test", {})
if str(availability.get("status", "")).strip().lower() == "unavailable":
return "blocked_unavailable"
if speed_test.get("tested_at") and speed_test.get("download_mbps") and speed_test.get("upload_mbps"):
return "post_install_verified"
if appointment.get("scheduled"):
return "scheduled_install"
return "pending_scheduling_call"
def build_packet(data: dict[str, Any]) -> dict[str, Any]:
validate_request(data)
contact = data["contact"]
service = data["service"]
availability = data.get("availability", {})
pricing = data.get("pricing", {})
appointment = data.get("appointment", {})
installer_access = data.get("installer_access", {})
payment = data.get("payment", {})
speed_test = data.get("speed_test", {})
packet = {
return {
"packet_id": f"nh-bb-{datetime.now(timezone.utc).strftime('%Y%m%d-%H%M%S')}",
"generated_utc": datetime.now(timezone.utc).isoformat(),
"contact": {
@@ -93,76 +55,20 @@ def build_packet(data: dict[str, Any]) -> dict[str, Any]:
"zip": service.get("zip", ""),
},
"desired_plan": data.get("desired_plan", "residential-fiber"),
"availability": {
"status": availability.get("status", "unknown"),
"checked_at": availability.get("checked_at", ""),
"notes": availability.get("notes", ""),
"exact_address_confirmed": bool(availability.get("exact_address_confirmed", False)),
},
"pricing": {
"recommended_plan": pricing.get("recommended_plan", data.get("desired_plan", "residential-fiber")),
"monthly_cost_usd": pricing.get("monthly_cost_usd"),
"install_fee_usd": pricing.get("install_fee_usd"),
"notes": pricing.get("notes", ""),
},
"appointment": {
"scheduled": bool(appointment.get("scheduled", False)),
"date": appointment.get("date", ""),
"window": appointment.get("window", ""),
"confirmation_number": appointment.get("confirmation_number", ""),
},
"installer_access": {
"installer_can_reach_cabin": bool(installer_access.get("installer_can_reach_cabin", False)),
"driveway_note": installer_access.get("driveway_note", ""),
"site_contact": installer_access.get("site_contact", contact["phone"]),
},
"payment": {
"method": payment.get("method", ""),
"first_month_due_usd": payment.get("first_month_due_usd"),
"install_fee_due_usd": payment.get("install_fee_due_usd"),
"notes": payment.get("notes", ""),
},
"speed_test": {
"tested_at": speed_test.get("tested_at", ""),
"download_mbps": speed_test.get("download_mbps"),
"upload_mbps": speed_test.get("upload_mbps"),
"provider": speed_test.get("provider", ""),
},
"call_log": data.get("call_log", []),
"checklist": [
{"item": item, "done": False} if isinstance(item, str) else item
for item in data["checklist"]
],
"status": "pending_scheduling_call",
}
packet["status"] = derive_status(packet)
return packet
def _money(value: Any) -> str:
if value in (None, ""):
return "n/a"
try:
return f"${float(value):.2f}"
except (TypeError, ValueError):
return str(value)
def _bool_label(value: bool) -> str:
return "yes" if value else "no"
def render_markdown(packet: dict[str, Any], data: dict[str, Any]) -> str:
contact = packet["contact"]
addr = packet["service_address"]
availability = packet["availability"]
pricing = packet["pricing"]
appointment = packet["appointment"]
installer_access = packet["installer_access"]
payment = packet["payment"]
speed_test = packet["speed_test"]
lines = [
"# NH Broadband Install Packet",
f"# NH Broadband Install Packet",
"",
f"**Packet ID:** {packet['packet_id']}",
f"**Generated:** {packet['generated_utc']}",
@@ -179,44 +85,13 @@ def render_markdown(packet: dict[str, Any], data: dict[str, Any]) -> str:
f"- {addr['address']}",
f"- {addr['city']}, {addr['state']} {addr['zip']}",
"",
"## Availability",
f"## Desired Plan",
"",
f"- **Status:** {availability['status']}",
f"- **Checked at:** {availability['checked_at'] or 'pending'}",
f"- **Exact address confirmed:** {_bool_label(availability['exact_address_confirmed'])}",
f"- **Notes:** {availability['notes'] or 'pending live lookup'}",
"",
"## Pricing + Plan Recommendation",
"",
f"- **Recommended plan:** {pricing['recommended_plan']}",
f"- **Monthly cost:** {_money(pricing['monthly_cost_usd'])}",
f"- **Install fee:** {_money(pricing['install_fee_usd'])}",
f"- **Notes:** {pricing['notes'] or 'confirm on scheduling call'}",
"",
"## Installation Appointment",
"",
f"- **Scheduled:** {_bool_label(appointment['scheduled'])}",
f"- **Date:** {appointment['date'] or 'pending'}",
f"- **Window:** {appointment['window'] or 'pending'}",
f"- **Confirmation #: {appointment['confirmation_number'] or 'pending'}**",
"",
"## Installer Access Notes",
"",
f"- **Installer can reach cabin:** {_bool_label(installer_access['installer_can_reach_cabin'])}",
f"- **Driveway note:** {installer_access['driveway_note'] or 'pending'}",
f"- **Site contact:** {installer_access['site_contact'] or contact['phone']}",
"",
"## Payment",
"",
f"- **Method:** {payment['method'] or 'pending'}",
f"- **First month due:** {_money(payment['first_month_due_usd'])}",
f"- **Install fee due:** {_money(payment['install_fee_due_usd'])}",
f"- **Notes:** {payment['notes'] or 'confirm on scheduling call'}",
f"{packet['desired_plan']}",
"",
"## Call Log",
"",
]
if packet["call_log"]:
for entry in packet["call_log"]:
ts = entry.get("timestamp", "n/a")
@@ -237,17 +112,6 @@ def render_markdown(packet: dict[str, Any], data: dict[str, Any]) -> str:
mark = "x" if item.get("done") else " "
lines.append(f"- [{mark}] {item['item']}")
if speed_test.get("tested_at") or speed_test.get("download_mbps") or speed_test.get("upload_mbps"):
lines.extend([
"",
"## Post-install Speed Test",
"",
f"- **Tested at:** {speed_test['tested_at'] or 'pending'}",
f"- **Download:** {speed_test['download_mbps'] or 'pending'} Mbps",
f"- **Upload:** {speed_test['upload_mbps'] or 'pending'} Mbps",
f"- **Provider:** {speed_test['provider'] or 'pending'}",
])
lines.append("")
return "\n".join(lines)

View File

@@ -0,0 +1,418 @@
#!/usr/bin/env python3
"""Ground timmy-home #524 as an executable status report.
Refs: timmy-home #524
"""
from __future__ import annotations
import argparse
import json
from copy import deepcopy
from pathlib import Path
from typing import Any
from urllib import request
DEFAULT_BASE_URL = "https://forge.alexanderwhitestone.com/api/v1"
DEFAULT_OWNER = "Timmy_Foundation"
DEFAULT_REPO = "timmy-home"
DEFAULT_TOKEN_FILE = Path.home() / ".config" / "gitea" / "token"
DEFAULT_REPO_ROOT = Path(__file__).resolve().parents[1]
DEFAULT_DOC_PATH = DEFAULT_REPO_ROOT / "docs" / "UNIFIED_FLEET_SOVEREIGNTY_STATUS.md"
DIRECTIVE_TITLE = "[DIRECTIVE] Unified Fleet Sovereignty & Comms Migration"
DIRECTIVE_SUMMARY = (
"Issue #524 is a multi-lane directive, not a one-commit feature. "
"This report grounds the directive in repo evidence, highlights stale cross-links, "
"and names the missing operator bundles that still need real execution."
)
DEFAULT_REFERENCE_SNAPSHOT = {
388: {
"title": "[KT] Fleet Lexicon & Techniques — Shared Vocabulary, Patterns, and Standards for All Agents",
"state": "closed",
},
103: {
"title": "Build comprehensive caching layer — cache everywhere",
"state": "open",
},
139: {
"title": "🐣 Allegro-Primus is born",
"state": "open",
},
813: {
"title": "docs: refresh the-playground genome analysis (#671)",
"state": "closed",
},
819: {
"title": "docs: verify #648 already implemented (closes #818)",
"state": "open",
},
}
WORKSTREAMS = [
{
"key": "nostr-migration",
"name": "Nostr Migration Leadership",
"requirement": "Replace Telegram with relay-based sovereign comms, verify wizard keypairs, and prove the NIP-29 group path is stable.",
"references": [813, 819],
"expected_keywords": ["nostr", "relay", "telegram", "comms", "messenger"],
"repo_evidence": [
{
"path": "infrastructure/timmy-bridge/client/timmy_client.py",
"description": "Nostr event client scaffold already exists",
},
{
"path": "infrastructure/timmy-bridge/monitor/timmy_monitor.py",
"description": "Nostr relay monitor already exists",
},
{
"path": "specs/wizard-telegram-bot-cutover.md",
"description": "Telegram cutover planning exists, so the migration lane is real",
},
],
"missing_deliverables": [
"wizard keypair inventory and ownership matrix",
"NIP-29 relay group verification report",
"operator runbook for cutting traffic off Telegram",
],
"why_open": "The repo has Nostr-adjacent scaffolding, but the directive still lacks a verified migration packet and the cited issue links drift away from the stated Nostr scope.",
},
{
"key": "lexicon-enforcement",
"name": "Lexicon Enforcement",
"requirement": "Enforce the Fleet Lexicon in PR review and issue triage so the team uses one shared language.",
"references": [388],
"expected_keywords": ["lexicon", "vocabulary", "standards", "shared vocabulary"],
"repo_evidence": [
{
"path": "docs/WIZARD_APPRENTICESHIP_CHARTER.md",
"description": "The repo already uses wizard-language canon in docs",
},
{
"path": "specs/timmy-ezra-bezalel-canon-sheet.md",
"description": "Canonical agent naming already exists",
},
{
"path": "docs/OPERATIONS_DASHBOARD.md",
"description": "Operational roles are already described in repo language",
},
],
"missing_deliverables": [
"machine-checkable lexicon policy for review/triage",
"terminology lint or reviewer checklist tied to the lexicon",
],
"why_open": "The naming canon exists, but there is still no executable enforcement bundle that would catch drift during future reviews and triage passes.",
},
{
"key": "feature-audit",
"name": "v0.7.0 Feature Audit",
"requirement": "Audit Hermes features that can reduce cloud dependency and turn the findings into a sovereignty implementation plan.",
"references": [139],
"expected_keywords": ["hermes", "feature", "audit", "v0.7.0", "sovereignty"],
"repo_evidence": [
{
"path": "scripts/sovereignty_audit.py",
"description": "Cloud-vs-local audit machinery already exists",
},
{
"path": "reports/evaluations/2026-04-15-phase-4-sovereignty-audit.md",
"description": "Recent sovereignty audit report is committed",
},
{
"path": "timmy-local/README.md",
"description": "Local-first status is already documented for operators",
},
],
"missing_deliverables": [
"Hermes v0.7.0 feature inventory linked to cloud-reduction leverage",
"Sovereignty Implementation Plan derived from that feature audit",
],
"why_open": "The repo has sovereignty-audit infrastructure, but it does not yet contain the requested v0.7.0 feature inventory or the plan that turns those findings into rollout steps.",
},
{
"key": "morrowind-benchmark",
"name": "Morrowind Local-First Benchmark",
"requirement": "Compare cloud and local Morrowind agents, prove local parity where possible, and document the reasoning gap when it fails.",
"references": [103],
"expected_keywords": ["morrowind", "combat", "benchmark", "local", "cloud"],
"repo_evidence": [
{
"path": "morrowind/local_brain.py",
"description": "Local Morrowind control loop already exists",
},
{
"path": "morrowind/mcp_server.py",
"description": "Morrowind MCP control surface is already wired",
},
{
"path": "morrowind/pilot.py",
"description": "Trajectory logging for evaluation already exists",
},
],
"missing_deliverables": [
"cloud-vs-local benchmark report for the combat loop",
"reasoning-gap writeup tied to a proposed LoRA/fine-tune path",
],
"why_open": "The repo has a local Morrowind stack, but it does not yet contain the requested benchmark artifact; the cited issue number also points at an unrelated caching task.",
},
{
"key": "syntax-guard",
"name": "Infrastructure Hardening / Syntax Guard",
"requirement": "Verify Syntax Guard pre-receive protection across Gitea repos so syntax failures stop earlier.",
"references": [],
"expected_keywords": [],
"repo_evidence": [],
"missing_deliverables": [
"repo inventory of Gitea targets that should carry Syntax Guard",
"deployment verifier for hook presence across those repos",
"operator report proving installation state instead of assuming it",
],
"why_open": "No repo-managed syntax-guard verifier is present yet, so this directive still depends on manual trust rather than auditable proof.",
},
]
def default_snapshot() -> dict[int, dict[str, str]]:
return deepcopy(DEFAULT_REFERENCE_SNAPSHOT)
class GiteaClient:
def __init__(self, token: str, owner: str = DEFAULT_OWNER, repo: str = DEFAULT_REPO, base_url: str = DEFAULT_BASE_URL):
self.token = token
self.owner = owner
self.repo = repo
self.base_url = base_url.rstrip("/")
def get_issue(self, issue_number: int) -> dict[str, Any]:
req = request.Request(
f"{self.base_url}/repos/{self.owner}/{self.repo}/issues/{issue_number}",
headers={"Authorization": f"token {self.token}", "Accept": "application/json"},
)
with request.urlopen(req, timeout=30) as resp:
return json.loads(resp.read().decode())
def load_snapshot(path: Path | None = None) -> dict[int, dict[str, str]]:
if path is None:
return default_snapshot()
data = json.loads(path.read_text(encoding="utf-8"))
return {int(k): v for k, v in data.items()}
def refresh_snapshot(token_file: Path = DEFAULT_TOKEN_FILE) -> dict[int, dict[str, str]]:
token = token_file.read_text(encoding="utf-8").strip()
client = GiteaClient(token=token)
snapshot: dict[int, dict[str, str]] = {}
for issue_number in sorted(DEFAULT_REFERENCE_SNAPSHOT):
issue = client.get_issue(issue_number)
snapshot[issue_number] = {
"title": issue["title"],
"state": issue["state"],
}
return snapshot
def collect_repo_evidence(entries: list[dict[str, str]], repo_root: Path) -> tuple[list[str], list[str]]:
present: list[str] = []
missing: list[str] = []
for entry in entries:
label = f"`{entry['path']}` — {entry['description']}"
if (repo_root / entry["path"]).exists():
present.append(label)
else:
missing.append(label)
return present, missing
def evaluate_reference(issue_number: int, snapshot: dict[int, dict[str, str]], expected_keywords: list[str]) -> dict[str, Any]:
record = snapshot.get(issue_number, {"title": "missing from snapshot", "state": "unknown"})
title = record["title"]
title_lower = title.lower()
matched_keywords = [kw for kw in expected_keywords if kw.lower() in title_lower]
aligned = bool(matched_keywords) if expected_keywords else True
return {
"number": issue_number,
"title": title,
"state": record["state"],
"aligned": aligned,
"matched_keywords": matched_keywords,
}
def classify_workstream(reference_results: list[dict[str, Any]], evidence_present: list[str], missing_deliverables: list[str]) -> str:
has_drift = any(not item["aligned"] for item in reference_results)
if not evidence_present:
return "MISSING"
if has_drift or missing_deliverables:
return "PARTIAL"
return "GROUNDED"
def evaluate_directive(snapshot: dict[int, dict[str, str]] | None = None, repo_root: Path | None = None) -> dict[str, Any]:
snapshot = snapshot or default_snapshot()
repo_root = repo_root or DEFAULT_REPO_ROOT
workstreams: list[dict[str, Any]] = []
drift_items: list[str] = []
for lane in WORKSTREAMS:
reference_results = [
evaluate_reference(issue_number, snapshot, lane["expected_keywords"])
for issue_number in lane["references"]
]
present, missing = collect_repo_evidence(lane["repo_evidence"], repo_root)
for item in reference_results:
if not item["aligned"]:
drift_items.append(
f"#{item['number']} is cited for {lane['name']}, but its current title is '{item['title']}'."
)
workstream = {
"key": lane["key"],
"name": lane["name"],
"requirement": lane["requirement"],
"reference_results": reference_results,
"repo_evidence_present": present,
"repo_evidence_missing": missing,
"missing_deliverables": list(lane["missing_deliverables"]),
"why_open": lane["why_open"],
}
workstream["status"] = classify_workstream(
reference_results=reference_results,
evidence_present=present,
missing_deliverables=workstream["missing_deliverables"],
)
workstreams.append(workstream)
next_actions: list[str] = []
for workstream in workstreams:
if workstream["missing_deliverables"]:
next_actions.append(f"{workstream['name']}: {workstream['missing_deliverables'][0]}")
return {
"issue_number": 524,
"title": DIRECTIVE_TITLE,
"summary": DIRECTIVE_SUMMARY,
"reference_snapshot": {str(k): v for k, v in sorted(snapshot.items())},
"workstreams": workstreams,
"reference_drift": drift_items,
"grounded_workstreams": sum(1 for item in workstreams if item["status"] == "GROUNDED"),
"partial_workstreams": sum(1 for item in workstreams if item["status"] == "PARTIAL"),
"missing_workstreams": sum(1 for item in workstreams if item["status"] == "MISSING"),
"next_actions": next_actions,
}
def render_markdown(result: dict[str, Any]) -> str:
lines = [
f"# {result['title']}",
"",
"Grounding report for `timmy-home #524`.",
"",
result["summary"],
"",
"This remains a `Refs #524` artifact. The directive spans multiple repos and operator actions, so this report makes the current repo-side state executable without pretending the whole migration is complete.",
"",
"## Directive Snapshot",
"",
f"- Repo-grounded workstreams: {result['grounded_workstreams']}",
f"- Partial workstreams: {result['partial_workstreams']}",
f"- Missing workstreams: {result['missing_workstreams']}",
f"- Drifted references: {len(result['reference_drift'])}",
"",
"## Reference Drift",
"",
]
if result["reference_drift"]:
lines.extend(f"- {item}" for item in result["reference_drift"])
else:
lines.append("- No stale cross-links detected in the directive snapshot.")
lines.extend(["", "## Workstream Matrix", ""])
for index, workstream in enumerate(result["workstreams"], start=1):
lines.extend(
[
f"### {index}. {workstream['name']}{workstream['status']}",
"",
f"- Requirement: {workstream['requirement']}",
]
)
if workstream["reference_results"]:
lines.append("- Referenced issues:")
for ref in workstream["reference_results"]:
alignment = "aligned" if ref["aligned"] else "DRIFT"
lines.append(
f" - #{ref['number']} ({ref['state']}) — {ref['title']} [{alignment}]"
)
else:
lines.append("- Referenced issues: none listed in the directive body")
if workstream["repo_evidence_present"]:
lines.append("- Repo evidence present:")
lines.extend(f" - {item}" for item in workstream["repo_evidence_present"])
else:
lines.append("- Repo evidence present: none")
if workstream["repo_evidence_missing"]:
lines.append("- Repo evidence expected but missing:")
lines.extend(f" - {item}" for item in workstream["repo_evidence_missing"])
if workstream["missing_deliverables"]:
lines.append("- Missing operator deliverables:")
lines.extend(f" - {item}" for item in workstream["missing_deliverables"])
else:
lines.append("- Missing operator deliverables: none")
lines.append(f"- Why this lane remains open: {workstream['why_open']}")
lines.append("")
lines.extend(["## Highest-Leverage Next Actions", ""])
lines.extend(f"- {item}" for item in result["next_actions"])
lines.extend(
[
"",
"## Why #524 Remains Open",
"",
"- The directive bundles five separate workstreams with different evidence surfaces.",
"- Multiple cited issue numbers have drifted away from the work they are supposed to anchor.",
"- Repo scaffolding exists for Nostr, sovereignty audits, and Morrowind, but the operator-facing bundles are still missing.",
"- Syntax Guard verification is still undocumented and unproven inside this repo.",
]
)
return "\n".join(lines).rstrip() + "\n"
def main() -> None:
parser = argparse.ArgumentParser(description="Render the unified fleet sovereignty status report for issue #524")
parser.add_argument("--snapshot", help="Optional JSON snapshot file overriding the default issue-title/state snapshot")
parser.add_argument("--live", action="store_true", help="Refresh the issue snapshot from Gitea before rendering")
parser.add_argument("--token-file", default=str(DEFAULT_TOKEN_FILE), help="Token file used with --live")
parser.add_argument("--output", help="Optional path to write the rendered report")
parser.add_argument("--json", action="store_true", help="Print computed JSON instead of markdown")
args = parser.parse_args()
if args.live:
snapshot = refresh_snapshot(Path(args.token_file).expanduser())
else:
snapshot = load_snapshot(Path(args.snapshot).expanduser() if args.snapshot else None)
result = evaluate_directive(snapshot=snapshot, repo_root=DEFAULT_REPO_ROOT)
rendered = json.dumps(result, indent=2) if args.json else render_markdown(result)
if args.output:
output_path = Path(args.output).expanduser()
output_path.parent.mkdir(parents=True, exist_ok=True)
output_path.write_text(rendered, encoding="utf-8")
print(f"Directive status written to {output_path}")
else:
print(rendered)
if __name__ == "__main__":
main()

View File

@@ -32,45 +32,11 @@ def test_load_and_build_packet() -> None:
assert packet["contact"]["name"] == "Timmy Operator"
assert packet["service_address"]["city"] == "Concord"
assert packet["service_address"]["state"] == "NH"
assert packet["availability"]["status"] == "available"
assert packet["appointment"]["scheduled"] is True
assert packet["pricing"]["monthly_cost_usd"] == 79.95
assert packet["installer_access"]["installer_can_reach_cabin"] is True
assert packet["payment"]["method"] == "credit_card"
assert packet["status"] == "scheduled_install"
assert packet["status"] == "pending_scheduling_call"
assert len(packet["checklist"]) == 8
assert packet["checklist"][0]["done"] is False
def test_build_packet_marks_blocked_when_availability_fails() -> None:
data = load_request("docs/nh-broadband-install-request.example.yaml")
data["availability"] = {
"status": "unavailable",
"checked_at": "2026-04-17T16:00:00Z",
"notes": "Address lookup returned no fiber service.",
}
data["appointment"] = {}
data["speed_test"] = {}
packet = build_packet(data)
assert packet["status"] == "blocked_unavailable"
def test_build_packet_marks_post_install_verified_when_speed_test_present() -> None:
data = load_request("docs/nh-broadband-install-request.example.yaml")
data["speed_test"] = {
"tested_at": "2026-05-01T18:30:00Z",
"download_mbps": 942.6,
"upload_mbps": 881.4,
"provider": "fast.com",
}
packet = build_packet(data)
assert packet["status"] == "post_install_verified"
def test_validate_rejects_missing_contact_name() -> None:
data = {
"contact": {"name": "", "phone": "555"},
@@ -120,11 +86,6 @@ def test_render_markdown_contains_key_sections() -> None:
assert "# NH Broadband Install Packet" in md
assert "## Contact" in md
assert "## Service Address" in md
assert "## Availability" in md
assert "## Pricing + Plan Recommendation" in md
assert "## Installation Appointment" in md
assert "## Installer Access Notes" in md
assert "## Payment" in md
assert "## Call Log" in md
assert "## Appointment Checklist" in md
assert "Concord" in md
@@ -136,8 +97,6 @@ def test_render_markdown_shows_checklist_items() -> None:
packet = build_packet(data)
md = render_markdown(packet, data)
assert "- [ ] Confirm exact-address availability" in md
assert "Installer can reach cabin" in md
assert "- **Confirmation #: NHB-2026-0417**" in md
def test_example_yaml_is_valid() -> None:

View File

@@ -0,0 +1,77 @@
from __future__ import annotations
import importlib.util
from pathlib import Path
ROOT = Path(__file__).resolve().parents[1]
SCRIPT_PATH = ROOT / "scripts" / "unified_fleet_sovereignty_status.py"
DOC_PATH = ROOT / "docs" / "UNIFIED_FLEET_SOVEREIGNTY_STATUS.md"
def _load_module(path: Path, name: str):
assert path.exists(), f"missing {path.relative_to(ROOT)}"
spec = importlib.util.spec_from_file_location(name, path)
assert spec and spec.loader
module = importlib.util.module_from_spec(spec)
spec.loader.exec_module(module)
return module
def _workstream(result: dict, key: str) -> dict:
for workstream in result["workstreams"]:
if workstream["key"] == key:
return workstream
raise AssertionError(f"missing workstream {key}")
def test_evaluate_directive_flags_reference_drift_without_faking_completion() -> None:
mod = _load_module(SCRIPT_PATH, "unified_fleet_sovereignty_status")
result = mod.evaluate_directive(snapshot=mod.default_snapshot(), repo_root=ROOT)
assert len(result["reference_drift"]) == 4
assert any("#813" in item for item in result["reference_drift"])
assert any("#103" in item for item in result["reference_drift"])
nostr = _workstream(result, "nostr-migration")
assert nostr["status"] == "PARTIAL"
assert any("timmy_client.py" in item for item in nostr["repo_evidence_present"])
lexicon = _workstream(result, "lexicon-enforcement")
assert all(item["aligned"] for item in lexicon["reference_results"])
assert lexicon["status"] == "PARTIAL"
syntax_guard = _workstream(result, "syntax-guard")
assert syntax_guard["status"] == "MISSING"
assert any("deployment verifier" in item for item in syntax_guard["missing_deliverables"])
def test_render_markdown_includes_required_sections_and_grounding_evidence() -> None:
mod = _load_module(SCRIPT_PATH, "unified_fleet_sovereignty_status")
result = mod.evaluate_directive(snapshot=mod.default_snapshot(), repo_root=ROOT)
report = mod.render_markdown(result)
for snippet in (
"# [DIRECTIVE] Unified Fleet Sovereignty & Comms Migration",
"## Directive Snapshot",
"## Reference Drift",
"## Workstream Matrix",
"### 5. Infrastructure Hardening / Syntax Guard — MISSING",
"`infrastructure/timmy-bridge/client/timmy_client.py`",
"machine-checkable lexicon policy for review/triage",
"## Why #524 Remains Open",
):
assert snippet in report
def test_repo_contains_committed_issue_524_grounding_doc() -> None:
assert DOC_PATH.exists(), "missing committed directive grounding doc"
text = DOC_PATH.read_text(encoding="utf-8")
for snippet in (
"# [DIRECTIVE] Unified Fleet Sovereignty & Comms Migration",
"## Reference Drift",
"## Workstream Matrix",
"## Highest-Leverage Next Actions",
"## Why #524 Remains Open",
):
assert snippet in text