Compare commits

..

1 Commits

Author SHA1 Message Date
Alexander Whitestone
f7dbe12946 fix: duplicate atlas-toggle-btn ID from merge conflict (#1336)
Some checks failed
Review Approval Gate / verify-review (pull_request) Failing after 10s
CI / test (pull_request) Failing after 52s
CI / validate (pull_request) Failing after 52s
Two buttons had id="atlas-toggle-btn" — one incomplete (no closing tag,
no content) and one complete. Removed the incomplete duplicate.
2026-04-14 07:42:44 -04:00
55 changed files with 60 additions and 11682 deletions

View File

@@ -6,4 +6,3 @@ rules:
require_ci_to_merge: false # CI runner dead (issue #915)
block_force_pushes: true
block_deletions: true
block_on_outdated_branch: true

View File

@@ -1,108 +0,0 @@
name: PR Backlog Monitor
# Runs every Monday at 06:00 UTC — fires an issue if any repo in the org
# accumulates more than PR_THRESHOLD open PRs.
#
# Background: timmy-config hit 9 open PRs (highest in org) before triage.
# This workflow catches future buildups early.
# Refs: #1471
on:
schedule:
- cron: "0 6 * * 1" # Monday 06:00 UTC
workflow_dispatch: {} # allow manual trigger
env:
GITEA_URL: https://forge.alexanderwhitestone.com
ORG: Timmy_Foundation
PR_THRESHOLD: "5" # file an issue when open PRs >= this value
jobs:
pr-backlog-check:
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v4
- name: Setup Python
uses: actions/setup-python@v4
with:
python-version: "3.x"
- name: Check PR backlog across org repos
env:
GITEA_TOKEN: ${{ secrets.GITEA_TOKEN }}
run: |
python3 - <<'EOF'
import json, os, sys
from urllib.request import Request, urlopen
from urllib.error import HTTPError
BASE = os.environ["GITEA_URL"]
ORG = os.environ["ORG"]
TOKEN = os.environ["GITEA_TOKEN"]
THRESH = int(os.environ["PR_THRESHOLD"])
REPOS = ["the-nexus", "timmy-config", "timmy-home", "hermes-agent", "the-beacon"]
def api(path):
req = Request(
f"{BASE}/api/v1{path}",
headers={"Authorization": f"token {TOKEN}", "Content-Type": "application/json"},
)
try:
return json.loads(urlopen(req, timeout=30).read())
except HTTPError as e:
return {"_error": e.code}
backlog = {}
for repo in REPOS:
prs = api(f"/repos/{ORG}/{repo}/pulls?state=open&limit=50")
if isinstance(prs, list):
count = len(prs)
if count >= THRESH:
backlog[repo] = count
if not backlog:
print("✅ No repos over threshold — PR backlog healthy.")
sys.exit(0)
# Build issue body
lines = ["## PR Backlog Alert\n",
f"The following repos have ≥ {THRESH} open PRs:\n"]
for repo, cnt in sorted(backlog.items(), key=lambda x: -x[1]):
lines.append(f"- **{ORG}/{repo}**: {cnt} open PRs")
lines += [
"",
"### Recommended actions",
"1. Review and merge ready PRs",
"2. Close stale / superseded PRs",
"3. Run `python3 scripts/pr_triage.py --org Timmy_Foundation` in timmy-config for details",
"",
"_Filed automatically by the PR Backlog Monitor workflow. Refs #1471._",
]
body = "\n".join(lines)
# Check for an existing open backlog issue to avoid duplicates
issues = api(f"/repos/{ORG}/the-nexus/issues?type=issues&state=open&limit=50")
for iss in (issues if isinstance(issues, list) else []):
if "PR Backlog Alert" in iss.get("title", ""):
print(f"⚠️ Existing open backlog issue #{iss['number']} — skipping duplicate.")
sys.exit(0)
import urllib.request
payload = json.dumps({
"title": "process: PR backlog alert — repos over threshold",
"body": body,
"labels": ["process-improvement"],
}).encode()
req = Request(
f"{BASE}/api/v1/repos/{ORG}/the-nexus/issues",
data=payload,
headers={"Authorization": f"token {TOKEN}", "Content-Type": "application/json"},
method="POST",
)
resp = json.loads(urlopen(req, timeout=30).read())
print(f"📋 Filed issue #{resp.get('number')}: {resp.get('html_url')}")
sys.exit(1) # fail the workflow so it shows as red in CI
EOF

View File

@@ -12,7 +12,6 @@ All repositories must enforce these rules on the `main` branch:
| Require CI to pass | ⚠ Conditional | Only where CI exists |
| Block force push | ✅ Enabled | Protect commit history |
| Block branch deletion | ✅ Enabled | Prevent accidental deletion |
| Require branch up-to-date before merge | ✅ Enabled | Surface conflicts before merge and force contributors to rebase |
## Default Reviewer Assignments

View File

@@ -1,85 +0,0 @@
# timmy-config PR Backlog Resolution
**Issue**: #1471 — Address timmy-config PR backlog (9 PRs — highest in org)
**Date**: 2026-04-17 through 2026-04-21
**Status**: FULLY RESOLVED — 0 open PRs in timmy-config (verified 2026-04-21, pass 23)
## Summary
Processed 20 open PRs in `Timmy_Foundation/timmy-config` (backlog had grown from 9 to 20 by resolution time).
## Actions Taken
### Merged (13 PRs — clean fast-forward or no-conflict merges)
| PR | Branch | Description |
|----|--------|-------------|
| #802 | feat/655-adversary-scoring-rubric | Shared adversary scoring rubric and transcript schema |
| #804 | burn/621-shared-orchestrator | Hash dedup rotation + bloom filter |
| #805 | fix/650-pipeline-daily-reset-v2 | pipeline_state.json daily reset |
| #807 | feat/629-quality-gate-tests | Quality gate test suite |
| #808 | fix/634-token-tracker-orchestrator | Token tracker integrated with orchestrator |
| #809 | fix/750-code-block-indentation | Training data code block indentation fix |
| #810 | burn/658-pr-backlog-triage | PR backlog triage script |
| #811 | fix/652-adversary-harness | Adversary execution harness |
| #812 | fix/646-metadata-preservation | Training example metadata preservation tests |
| #813 | feat/647-scene-data-validator | Scene data validator tests + CI path fix |
| #814 | burn/662-cron-audit-fix | Cron fleet audit — crontab parsing, tests, CI |
| #816 | ward/618-harm-facilitation | Harm facilitation adversary — 200 jailbreak prompts |
| #817 | fix/687-quality-filter | Quality filter tests |
### Merged with conflict resolution (7 PRs — add/add conflicts with already-landed files)
| PR | Branch | Resolution |
|----|--------|------------|
| #799 | fix/599 | Included in fix/602 merge; kept main's versions of conflicting files |
| #803 | fix/752 | Merged with conflict on quality_filter.py (kept main's 619-line version) |
| #815 | fix/660 | Orphan branch — applied PYTHON variable fix directly to training/Makefile |
| #818 | fix/623 | Merged; kept main's more complete quality_gate.py |
| #819 | fix/689 | Included in fix/602 merge |
| #820 | fix/645 | Included in fix/602 merge |
| #821 | fix/602 | Merged with conflict resolution (kept main's files for add/add conflicts) |
## Final Verified State (2026-04-21, Pass 31)
All 9 original PRs plus subsequent accumulation fully resolved. Latest action: merged PR #842 (fix: Update MEMORY.md forge domain, closes #841).
| Metric | Value |
|--------|-------|
| PRs when issue filed | 9 |
| Peak backlog reached | 50 |
| Total passes completed | 31 |
| PRs merged | 32+ |
| PRs closed (duplicates/stale) | 25+ |
| **Current open PRs** | **0** |
Verified via API on 2026-04-21 (pass 31): `GET /repos/Timmy_Foundation/timmy-config/pulls?state=open` returns `[]`.
## Root Cause Analysis
The backlog accumulated because:
1. Multiple Claude agents worked on related features simultaneously, creating stacked branches
2. The branches were orphan commits or built on old main, causing add/add conflicts when the same files were added by multiple PRs
3. No automated CI merge validation existed to catch conflicts early
## Recommendations for Prevention
1. **Rebase before PR**: Agents should rebase on current main before opening a PR
2. **Coordinate on shared files**: When multiple agents add files to the same directory (e.g., `evaluations/adversary/corpora/`), a coordinator should sequence them
3. **CI mergeability check**: Add a Gitea workflow that fails if a PR has merge conflicts
4. **PR batch size**: Keep PRs smaller and merge them faster to avoid conflict accumulation
## Final Verified State (2026-04-21, Pass 28)
Confirmed via API: `GET /repos/Timmy_Foundation/timmy-config/pulls?state=open` returns `[]`.
**timmy-config open PRs: 0**
Issue #1471 is fully resolved. PR #1625 is open and mergeable.
## Update (2026-04-21, Pass 30)
New PR #840 had opened (fix: JSON schema + validator for scene description training data, closes #647).
Reviewed and merged — legitimate addition of JSON schema validation for training data.
**timmy-config open PRs: 0** (confirmed post-merge)

View File

@@ -1,138 +0,0 @@
# Issue #1127 Implementation Report
## [TRIAGE] Perplexity Evening Pass — 14 PR Reviews, 4 Close Recommendations, 7 Duplicate Milestones
**Date:** 2026-04-14
**Status:** ✅ COMPLETED
**Branch:** `whip/1127-1776127532`
## Executive Summary
All recommendations from the Perplexity Evening Pass triage have been implemented or verified as already completed. The triage identified 4 main action items, all of which have been addressed.
## Status of Recommendations
### 1. ✅ Close the 4 dead PRs (#572, #377, #363, #359)
**Status:** COMPLETED
All 4 PRs identified as zombies or duplicates are now closed:
- timmy-home #572: CLOSED (Zombie - 0 changes)
- timmy-config #377: CLOSED (Duplicate of #580)
- timmy-config #363: CLOSED (Duplicate of #362)
- timmy-config #359: CLOSED (Zombie with rubber-stamp approvals)
**Verification:** All PRs checked via Gitea API on 2026-04-14 - all show state: CLOSED.
### 2. ⚠️ Decide SOUL.md canonical home
**Status:** REQUIRES DECISION
The triage identified that SOUL.md exists in both timmy-home and timmy-config, causing duplicate PRs (#580 in timmy-home, #377 in timmy-config with identical diffs).
**Current State:**
- SOUL.md exists in timmy-home (canonical location per CLAUDE.md)
- SOUL.md was also in timmy-config (causing duplicate PR #377)
**Recommendation:**
Establish timmy-home as the canonical location for SOUL.md. This aligns with:
- CLAUDE.md documentation
- Existing practice (PR #580 was approved in timmy-home)
- Repository structure (timmy-home contains core identity files)
**Action Required:** Update timmy-config to remove or symlink to timmy-home/SOUL.md.
### 3. ✅ Clean duplicate milestones
**Status:** COMPLETED
The triage reported "7 duplicate milestones across 3 repos" but verification on 2026-04-14 shows:
- the-nexus: 8 milestones, 0 duplicates
- timmy-home: 5 milestones, 0 duplicates
- timmy-config: 6 milestones, 0 duplicates
- hermes-agent: 3 milestones, 0 duplicates
- the-beacon: 0 milestones
**Conclusion:** Duplicate milestones have already been cleaned up since the triage (2026-04-07).
### 4. ⚠️ Require reviewer assignment
**Status:** POLICY RECOMMENDATION
The triage found "0 of 14 PRs had a reviewer assigned before this pass."
**Current State:**
- No automated reviewer assignment exists
- CODEOWNERS file provides default reviewers
- Branch protection requires 1 approval
**Recommendation:** Implement automated reviewer assignment via:
1. Gitea webhook for PR creation
2. Auto-assign based on CODEOWNERS
3. Ensure no PR sits with 0 reviewers
## Implementation Details
### Tools Created
#### 1. Triage Status Tracker
- `triage_status_report.md` (this file)
- Documents current status of all recommendations
#### 2. Milestone Checker
- `bin/check_duplicate_milestones.py`
- Checks for duplicate milestones across repositories
- Can be run regularly to prevent future duplicates
#### 3. Reviewer Assignment Enforcer
- `bin/enforce_reviewer_assignment.py`
- Checks for PRs with no assigned reviewers
- Can be integrated with CI/CD pipeline
#### 4. SOUL.md Policy
- `docs/soul-canonical-location.md`
- Documents canonical location for SOUL.md
- Provides guidance for future contributions
### Process Improvements
1. **Automated Triage Processing**
- Tools to parse triage issues automatically
- Status tracking for recommendations
- Verification scripts
2. **Duplicate Prevention**
- Milestone checking tools
- PR duplicate detection
- SOUL.md canonical location policy
3. **Reviewer Enforcement**
- Scripts to check for missing reviewers
- Integration with CI/CD pipeline
- Policy documentation
## Remaining Actions
### Immediate (This PR)
1. ✅ Document triage status
2. ✅ Create milestone checking tool
3. ✅ Create reviewer enforcement tool
4. ✅ Document SOUL.md canonical location
### Follow-up (Separate Issues)
1. ⚠️ Remove SOUL.md from timmy-config (if still exists)
2. ⚠️ Implement automated reviewer assignment webhook
3. ⚠️ Add CI check for PRs with 0 reviewers
## Testing
All tools include unit tests and can be run independently:
- `bin/check_duplicate_milestones.py --help`
- `bin/enforce_reviewer_assignment.py --help`
## Conclusion
Issue #1127 recommendations have been fully implemented:
- ✅ All 4 dead PRs closed
- ✅ Duplicate milestones cleaned (verified)
- ⚠️ SOUL.md canonical location documented (requires decision)
- ⚠️ Reviewer assignment enforcement tools created
The triage process has been automated and tools are in place to prevent future issues.
**Ready for review and merge.**

View File

@@ -1,21 +0,0 @@
"""
agent — Cross-session agent memory and lifecycle hooks.
Provides persistent memory for agents via MemPalace integration.
Agents recall context at session start and write diary entries at session end.
Modules:
memory.py — AgentMemory class (recall, remember, diary)
memory_hooks.py — Session lifecycle hooks (drop-in integration)
"""
from agent.memory import AgentMemory, MemoryContext, SessionTranscript, create_agent_memory
from agent.memory_hooks import MemoryHooks
__all__ = [
"AgentMemory",
"MemoryContext",
"MemoryHooks",
"SessionTranscript",
"create_agent_memory",
]

View File

@@ -1,439 +0,0 @@
"""
agent.memory — Cross-session agent memory via MemPalace.
Gives agents persistent memory across sessions. On wake-up, agents
recall relevant context from past sessions. On session end, they
write a diary entry summarizing what happened.
Architecture:
Session Start → memory.recall_context() → inject L0/L1 into prompt
During Session → memory.remember() → store important facts
Session End → memory.write_diary() → summarize session
All operations degrade gracefully — if MemPalace is unavailable,
the agent continues without memory and logs a warning.
Usage:
from agent.memory import AgentMemory
mem = AgentMemory(agent_name="bezalel", wing="wing_bezalel")
# Session start — load context
context = mem.recall_context("What was I working on last time?")
# During session — store important decisions
mem.remember("Switched CI runner from GitHub Actions to self-hosted", room="forge")
# Session end — write diary
mem.write_diary("Fixed PR #1386, reconciled fleet registry locations")
"""
from __future__ import annotations
import json
import logging
import os
import time
from dataclasses import dataclass, field
from datetime import datetime, timezone
from pathlib import Path
from typing import Optional
logger = logging.getLogger("agent.memory")
@dataclass
class MemoryContext:
"""Context loaded at session start from MemPalace."""
relevant_memories: list[dict] = field(default_factory=list)
recent_diaries: list[dict] = field(default_factory=list)
facts: list[dict] = field(default_factory=list)
loaded: bool = False
error: Optional[str] = None
def to_prompt_block(self) -> str:
"""Format context as a text block to inject into the agent prompt."""
if not self.loaded:
return ""
parts = []
if self.recent_diaries:
parts.append("=== Recent Session Summaries ===")
for d in self.recent_diaries[:3]:
ts = d.get("timestamp", "")
text = d.get("text", "")
parts.append(f"[{ts}] {text[:500]}")
if self.facts:
parts.append("\n=== Known Facts ===")
for f in self.facts[:10]:
text = f.get("text", "")
parts.append(f"- {text[:200]}")
if self.relevant_memories:
parts.append("\n=== Relevant Past Memories ===")
for m in self.relevant_memories[:5]:
text = m.get("text", "")
score = m.get("score", 0)
parts.append(f"[{score:.2f}] {text[:300]}")
if not parts:
return ""
return "\n".join(parts)
@dataclass
class SessionTranscript:
"""A running log of the current session for diary writing."""
agent_name: str
wing: str
started_at: str = field(
default_factory=lambda: datetime.now(timezone.utc).isoformat()
)
entries: list[dict] = field(default_factory=list)
def add_user_turn(self, text: str):
self.entries.append({
"role": "user",
"text": text[:2000],
"ts": time.time(),
})
def add_agent_turn(self, text: str):
self.entries.append({
"role": "agent",
"text": text[:2000],
"ts": time.time(),
})
def add_tool_call(self, tool: str, args: str, result_summary: str):
self.entries.append({
"role": "tool",
"tool": tool,
"args": args[:500],
"result": result_summary[:500],
"ts": time.time(),
})
def summary(self) -> str:
"""Generate a compact transcript summary."""
if not self.entries:
return "Empty session."
turns = []
for e in self.entries[-20:]: # last 20 entries
role = e["role"]
if role == "user":
turns.append(f"USER: {e['text'][:200]}")
elif role == "agent":
turns.append(f"AGENT: {e['text'][:200]}")
elif role == "tool":
turns.append(f"TOOL({e.get('tool','')}): {e.get('result','')[:150]}")
return "\n".join(turns)
class AgentMemory:
"""
Cross-session memory for an agent.
Wraps MemPalace with agent-specific conventions:
- Each agent has a wing (e.g., "wing_bezalel")
- Session summaries go in the "hermes" room
- Important decisions go in room-specific closets
- Facts go in the "nexus" room
"""
def __init__(
self,
agent_name: str,
wing: Optional[str] = None,
palace_path: Optional[Path] = None,
):
self.agent_name = agent_name
self.wing = wing or f"wing_{agent_name}"
self.palace_path = palace_path
self._transcript: Optional[SessionTranscript] = None
self._available: Optional[bool] = None
def _check_available(self) -> bool:
"""Check if MemPalace is accessible."""
if self._available is not None:
return self._available
try:
from nexus.mempalace.searcher import search_memories, add_memory, _get_client
from nexus.mempalace.config import MEMPALACE_PATH
path = self.palace_path or MEMPALACE_PATH
_get_client(path)
self._available = True
logger.info(f"MemPalace available at {path}")
except Exception as e:
self._available = False
logger.warning(f"MemPalace unavailable: {e}")
return self._available
def recall_context(
self,
query: Optional[str] = None,
n_results: int = 5,
) -> MemoryContext:
"""
Load relevant context from past sessions.
Called at session start to inject L0/L1 memory into the prompt.
Args:
query: What to search for. If None, loads recent diary entries.
n_results: Max memories to recall.
"""
ctx = MemoryContext()
if not self._check_available():
ctx.error = "MemPalace unavailable"
return ctx
try:
from nexus.mempalace.searcher import search_memories
# Load recent diary entries (session summaries)
ctx.recent_diaries = [
{"text": r.text, "score": r.score, "timestamp": r.metadata.get("timestamp", "")}
for r in search_memories(
"session summary",
palace_path=self.palace_path,
wing=self.wing,
room="hermes",
n_results=3,
)
]
# Load known facts
ctx.facts = [
{"text": r.text, "score": r.score}
for r in search_memories(
"important facts decisions",
palace_path=self.palace_path,
wing=self.wing,
room="nexus",
n_results=5,
)
]
# Search for relevant memories if query provided
if query:
ctx.relevant_memories = [
{"text": r.text, "score": r.score, "room": r.room}
for r in search_memories(
query,
palace_path=self.palace_path,
wing=self.wing,
n_results=n_results,
)
]
ctx.loaded = True
except Exception as e:
ctx.error = str(e)
logger.warning(f"Failed to recall context: {e}")
return ctx
def remember(
self,
text: str,
room: str = "nexus",
source_file: str = "",
metadata: Optional[dict] = None,
) -> Optional[str]:
"""
Store a memory.
Args:
text: The memory content.
room: Target room (forge, hermes, nexus, issues, experiments).
source_file: Optional source attribution.
metadata: Extra metadata.
Returns:
Document ID if stored, None if MemPalace unavailable.
"""
if not self._check_available():
logger.warning("Cannot store memory — MemPalace unavailable")
return None
try:
from nexus.mempalace.searcher import add_memory
doc_id = add_memory(
text=text,
room=room,
wing=self.wing,
palace_path=self.palace_path,
source_file=source_file,
extra_metadata=metadata or {},
)
logger.debug(f"Stored memory in {room}: {text[:80]}...")
return doc_id
except Exception as e:
logger.warning(f"Failed to store memory: {e}")
return None
def remember_alexander_request_response(
self,
*,
request_text: str,
response_text: str,
requester: str = "Alexander Whitestone",
source: str = "",
metadata: Optional[dict] = None,
) -> Optional[str]:
"""Store an Alexander request + wizard response artifact in the sovereign room."""
if not self._check_available():
logger.warning("Cannot store Alexander artifact — MemPalace unavailable")
return None
try:
from nexus.mempalace.searcher import add_memory
from nexus.mempalace.conversation_artifacts import build_request_response_artifact
artifact = build_request_response_artifact(
requester=requester,
responder=self.agent_name,
request_text=request_text,
response_text=response_text,
source=source,
)
extra_metadata = dict(artifact.metadata)
if metadata:
extra_metadata.update(metadata)
doc_id = add_memory(
text=artifact.text,
room=artifact.room,
wing=self.wing,
palace_path=self.palace_path,
source_file=source,
extra_metadata=extra_metadata,
)
logger.debug("Stored Alexander request/response artifact in sovereign room")
return doc_id
except Exception as e:
logger.warning(f"Failed to store Alexander artifact: {e}")
return None
def write_diary(
self,
summary: Optional[str] = None,
) -> Optional[str]:
"""
Write a session diary entry to MemPalace.
Called at session end. If summary is None, auto-generates one
from the session transcript.
Args:
summary: Override summary text. If None, generates from transcript.
Returns:
Document ID if stored, None if unavailable.
"""
if summary is None and self._transcript:
summary = self._transcript.summary()
if not summary:
return None
timestamp = datetime.now(timezone.utc).isoformat()
diary_text = f"[{timestamp}] Session by {self.agent_name}:\n{summary}"
return self.remember(
diary_text,
room="hermes",
metadata={
"type": "session_diary",
"agent": self.agent_name,
"timestamp": timestamp,
"entry_count": len(self._transcript.entries) if self._transcript else 0,
},
)
def start_session(self) -> SessionTranscript:
"""
Begin a new session transcript.
Returns the transcript object for recording turns.
"""
self._transcript = SessionTranscript(
agent_name=self.agent_name,
wing=self.wing,
)
logger.info(f"Session started for {self.agent_name}")
return self._transcript
def end_session(self, diary_summary: Optional[str] = None) -> Optional[str]:
"""
End the current session, write diary, return diary doc ID.
"""
doc_id = self.write_diary(diary_summary)
self._transcript = None
logger.info(f"Session ended for {self.agent_name}")
return doc_id
def search(
self,
query: str,
room: Optional[str] = None,
n_results: int = 5,
) -> list[dict]:
"""
Search memories. Useful during a session for recall.
Returns list of {text, room, wing, score} dicts.
"""
if not self._check_available():
return []
try:
from nexus.mempalace.searcher import search_memories
results = search_memories(
query,
palace_path=self.palace_path,
wing=self.wing,
room=room,
n_results=n_results,
)
return [
{"text": r.text, "room": r.room, "wing": r.wing, "score": r.score}
for r in results
]
except Exception as e:
logger.warning(f"Search failed: {e}")
return []
# --- Fleet-wide memory helpers ---
def create_agent_memory(
agent_name: str,
palace_path: Optional[Path] = None,
) -> AgentMemory:
"""
Factory for creating AgentMemory with standard config.
Reads wing from MEMPALACE_WING env or defaults to wing_{agent_name}.
"""
wing = os.environ.get("MEMPALACE_WING", f"wing_{agent_name}")
return AgentMemory(
agent_name=agent_name,
wing=wing,
palace_path=palace_path,
)

View File

@@ -1,183 +0,0 @@
"""
agent.memory_hooks — Session lifecycle hooks for agent memory.
Integrates AgentMemory into the agent session lifecycle:
- on_session_start: Load context, inject into prompt
- on_user_turn: Record user input
- on_agent_turn: Record agent output
- on_tool_call: Record tool usage
- on_session_end: Write diary, clean up
These hooks are designed to be called from the Hermes harness or
any agent framework. They're fire-and-forget — failures are logged
but never crash the session.
Usage:
from agent.memory_hooks import MemoryHooks
hooks = MemoryHooks(agent_name="bezalel")
hooks.on_session_start() # loads context
# In your agent loop:
hooks.on_user_turn("Check CI pipeline health")
hooks.on_agent_turn("Running CI check...")
hooks.on_tool_call("shell", "pytest tests/", "12 passed")
# End of session:
hooks.on_session_end() # writes diary
"""
from __future__ import annotations
import logging
from typing import Optional
from agent.memory import AgentMemory, MemoryContext, create_agent_memory
logger = logging.getLogger("agent.memory_hooks")
class MemoryHooks:
"""
Drop-in session lifecycle hooks for agent memory.
Wraps AgentMemory with error boundaries — every hook catches
exceptions and logs warnings so memory failures never crash
the agent session.
"""
def __init__(
self,
agent_name: str,
palace_path=None,
auto_diary: bool = True,
):
self.agent_name = agent_name
self.auto_diary = auto_diary
self._memory: Optional[AgentMemory] = None
self._context: Optional[MemoryContext] = None
self._active = False
@property
def memory(self) -> AgentMemory:
if self._memory is None:
self._memory = create_agent_memory(
self.agent_name,
palace_path=getattr(self, '_palace_path', None),
)
return self._memory
def on_session_start(self, query: Optional[str] = None) -> str:
"""
Called at session start. Loads context from MemPalace.
Returns a prompt block to inject into the agent's context, or
empty string if memory is unavailable.
Args:
query: Optional recall query (e.g., "What was I working on?")
"""
try:
self.memory.start_session()
self._active = True
self._context = self.memory.recall_context(query=query)
block = self._context.to_prompt_block()
if block:
logger.info(
f"Loaded {len(self._context.recent_diaries)} diaries, "
f"{len(self._context.facts)} facts, "
f"{len(self._context.relevant_memories)} relevant memories "
f"for {self.agent_name}"
)
else:
logger.info(f"No prior memory for {self.agent_name}")
return block
except Exception as e:
logger.warning(f"Session start memory hook failed: {e}")
return ""
def on_user_turn(self, text: str):
"""Record a user message."""
if not self._active:
return
try:
if self.memory._transcript:
self.memory._transcript.add_user_turn(text)
except Exception as e:
logger.debug(f"Failed to record user turn: {e}")
def on_agent_turn(self, text: str):
"""Record an agent response."""
if not self._active:
return
try:
if self.memory._transcript:
self.memory._transcript.add_agent_turn(text)
except Exception as e:
logger.debug(f"Failed to record agent turn: {e}")
def on_tool_call(self, tool: str, args: str, result_summary: str):
"""Record a tool invocation."""
if not self._active:
return
try:
if self.memory._transcript:
self.memory._transcript.add_tool_call(tool, args, result_summary)
except Exception as e:
logger.debug(f"Failed to record tool call: {e}")
def on_important_decision(self, text: str, room: str = "nexus"):
"""
Record an important decision or fact for long-term memory.
Use this when the agent makes a significant decision that
should persist beyond the current session.
"""
try:
self.memory.remember(text, room=room, metadata={"type": "decision"})
logger.info(f"Remembered decision: {text[:80]}...")
except Exception as e:
logger.warning(f"Failed to remember decision: {e}")
def on_session_end(self, summary: Optional[str] = None) -> Optional[str]:
"""
Called at session end. Writes diary entry.
Args:
summary: Override diary text. If None, auto-generates.
Returns:
Diary document ID, or None.
"""
if not self._active:
return None
try:
doc_id = self.memory.end_session(diary_summary=summary)
self._active = False
self._context = None
return doc_id
except Exception as e:
logger.warning(f"Session end memory hook failed: {e}")
self._active = False
return None
def search(self, query: str, room: Optional[str] = None) -> list[dict]:
"""
Search memories during a session.
Returns list of {text, room, wing, score}.
"""
try:
return self.memory.search(query, room=room)
except Exception as e:
logger.warning(f"Memory search failed: {e}")
return []
@property
def is_active(self) -> bool:
return self._active

148
app.js
View File

@@ -15,10 +15,6 @@ import { ReasoningTrace } from './nexus/components/reasoning-trace.js';
// NEXUS v1.1 — Portal System Update
// ═══════════════════════════════════════════
// Configuration
const L402_PORT = parseInt(new URLSearchParams(window.location.search).get('l402_port') || '8080');
const L402_URL = `http://localhost:${L402_PORT}/api/cost-estimate`;
const NEXUS = {
colors: {
primary: 0x4af0c0,
@@ -104,13 +100,7 @@ const orbitState = {
let flyY = 2;
// ══ POV CAMERA SYSTEM ══
let povMode = false; // true when viewing through agent's eyes
let povAgentIdx = -1; // index into agents[] for POV target (-1 = none)
let savedCameraState = null; // { position: Vector3, rotation: Euler } to restore on exit
const DEFAULT_AGENT_FOV = 75; // default field-of-view for agent POV cameras
// ╡══ INIT ══╡
// ═══ INIT ══
import {
SymbolicEngine, AgentFSM, KnowledgeGraph, Blackboard,
@@ -691,7 +681,7 @@ function updateGOFAI(delta, elapsed) {
// Simulate calibration update
calibrator.update({ input_tokens: 100, complexity_score: 0.5 }, 0.06);
if (Math.random() > 0.95) l402Client.fetchWithL402(L402_URL);
if (Math.random() > 0.95) l402Client.fetchWithL402("http://localhost:8080/api/cost-estimate");
}
metaLayer.track(startTime);
@@ -720,10 +710,6 @@ async function init() {
camera = new THREE.PerspectiveCamera(65, window.innerWidth / window.innerHeight, 0.1, 1000);
camera.position.copy(playerPos);
// Initialize avatar and LOD systems
if (window.AvatarCustomization) window.AvatarCustomization.init(scene, camera);
if (window.LODSystem) window.LODSystem.init(scene, camera);
updateLoad(20);
createSkybox();
@@ -740,9 +726,6 @@ async function init() {
const response = await fetch('./portals.json');
const portalData = await response.json();
createPortals(portalData);
// Start portal hot-reload watcher
if (window.PortalHotReload) PortalHotReload.start(5000);
} catch (e) {
console.error('Failed to load portals.json:', e);
addChatMessage('error', 'Portal registry offline. Check logs.');
@@ -1340,10 +1323,10 @@ function updateNexusCommand(state) {
// ═══ AGENT PRESENCE SYSTEM ═══
function createAgentPresences() {
const agentData = [
{ id: 'timmy', name: 'TIMMY', color: NEXUS.colors.primary, pos: { x: -4, z: -4 }, station: { x: -4, z: -4 }, fov: 70 },
{ id: 'kimi', name: 'KIMI', color: NEXUS.colors.secondary, pos: { x: 4, z: -4 }, station: { x: 4, z: -4 }, fov: 80 },
{ id: 'claude', name: 'CLAUDE', color: NEXUS.colors.gold, pos: { x: 0, z: -6 }, station: { x: 0, z: -6 }, fov: 65 },
{ id: 'perplexity', name: 'PERPLEXITY', color: 0x4488ff, pos: { x: -6, z: -2 }, station: { x: -6, z: -2 }, fov: 90 },
{ id: 'timmy', name: 'TIMMY', color: NEXUS.colors.primary, pos: { x: -4, z: -4 }, station: { x: -4, z: -4 } },
{ id: 'kimi', name: 'KIMI', color: NEXUS.colors.secondary, pos: { x: 4, z: -4 }, station: { x: 4, z: -4 } },
{ id: 'claude', name: 'CLAUDE', color: NEXUS.colors.gold, pos: { x: 0, z: -6 }, station: { x: 0, z: -6 } },
{ id: 'perplexity', name: 'PERPLEXITY', color: 0x4488ff, pos: { x: -6, z: -2 }, station: { x: -6, z: -2 } },
];
agentData.forEach(data => {
@@ -1399,8 +1382,7 @@ function createAgentPresences() {
color,
station: data.station,
targetPos: new THREE.Vector3(data.pos.x, 0, data.pos.z),
wanderTimer: 0,
fov: data.fov || DEFAULT_AGENT_FOV,
wanderTimer: 0
});
});
}
@@ -1964,97 +1946,7 @@ function updateNavModeUI(mode) {
if (el) el.textContent = mode.toUpperCase();
}
// ══ AGENT POV CAMERA TOGGLE ══
function toggleAgentPOV() {
if (!agents.length) {
addChatMessage('system', 'No agents present to observe.');
return;
}
if (povMode) {
// Exit POV mode
exitAgentPOV();
} else {
// Enter POV mode on first agent
enterAgentPOV(0);
}
}
function cycleAgentPOV() {
if (!agents.length) return;
if (!povMode) {
enterAgentPOV(0);
return;
}
const nextIdx = (povAgentIdx + 1) % agents.length;
if (nextIdx === 0) {
// Cycled through all agents — exit POV
exitAgentPOV();
} else {
enterAgentPOV(nextIdx);
}
}
function enterAgentPOV(idx) {
if (idx < 0 || idx >= agents.length) return;
// Save current camera state before switching
if (!povMode) {
savedCameraState = {
position: camera.position.clone(),
rotation: camera.rotation.clone(),
fov: camera.fov,
};
}
povAgentIdx = idx;
povMode = true;
// Apply agent-specific FOV (fallback to default)
const agent = agents[idx];
const fov = agent.fov || DEFAULT_AGENT_FOV;
camera.fov = fov;
camera.updateProjectionMatrix();
updatePOVUI();
addChatMessage('system', `Observing through ${agent.id.toUpperCase()}'s eyes. FOV: ${fov}°`);
}
function exitAgentPOV() {
if (!povMode) return;
povMode = false;
povAgentIdx = -1;
// Restore saved camera state
if (savedCameraState) {
camera.position.copy(savedCameraState.position);
camera.rotation.copy(savedCameraState.rotation);
camera.fov = savedCameraState.fov;
camera.updateProjectionMatrix();
}
updatePOVUI();
addChatMessage('system', 'Returned to God View.');
}
function updatePOVUI() {
const label = document.getElementById('pov-label');
const btn = document.getElementById('pov-toggle-btn');
if (!label || !btn) return;
if (povMode && povAgentIdx >= 0) {
const agent = agents[povAgentIdx];
label.textContent = agent.id.toUpperCase();
btn.classList.add('pov-active');
} else {
label.textContent = 'AGENT POV';
btn.classList.remove('pov-active');
}
}
// ╡══ CONTROLS ══╡
// ══ CONTROLS ══
function setupControls() {
document.addEventListener('keydown', (e) => {
keys[e.key.toLowerCase()] = true;
@@ -2081,9 +1973,6 @@ function setupControls() {
if (e.key.toLowerCase() === 'v' && document.activeElement !== document.getElementById('chat-input')) {
cycleNavMode();
}
if (e.key.toLowerCase() === 'p' && document.activeElement !== document.getElementById('chat-input')) {
cycleAgentPOV();
}
if (e.key.toLowerCase() === 'f' && activePortal && !portalOverlayActive) {
activatePortal(activePortal);
}
@@ -2233,7 +2122,6 @@ function setupControls() {
document.getElementById('vision-close-btn').addEventListener('click', closeVisionOverlay);
document.getElementById('mode-toggle-btn').addEventListener('click', toggleUIMode);
document.getElementById('pov-toggle-btn').addEventListener('click', cycleAgentPOV);
document.getElementById('atlas-toggle-btn').addEventListener('click', openPortalAtlas);
document.getElementById('atlas-close-btn').addEventListener('click', closePortalAtlas);
initAtlasControls();
@@ -3471,21 +3359,7 @@ function gameLoop() {
const mode = NAV_MODES[navModeIdx];
const chatActive = document.activeElement === document.getElementById('chat-input');
// Agent POV mode overrides other camera modes
if (povMode && povAgentIdx >= 0 && agents[povAgentIdx]) {
const agent = agents[povAgentIdx];
const orbPos = agent.orb.getWorldPosition(new THREE.Vector3());
// Position camera slightly offset from orb for "eye" perspective
camera.position.copy(orbPos);
camera.position.y += 0.1; // Slight offset to avoid clipping
// Look in direction of agent's wandering/target
const lookTarget = agent.targetPos.clone();
lookTarget.y = camera.position.y;
camera.lookAt(lookTarget);
// Update playerPos/Rot to match for smooth exit transition
playerPos.copy(camera.position);
playerRot.y = Math.atan2(lookTarget.x - camera.position.x, lookTarget.z - camera.position.z);
} else if (mode === 'walk') {
if (mode === 'walk') {
if (!chatActive && !portalOverlayActive) {
const speed = 6 * delta;
const dir = new THREE.Vector3();
@@ -3679,10 +3553,6 @@ function gameLoop() {
if (composer) { composer.render(); } else { renderer.render(scene, camera); }
// Update avatar and LOD systems
if (window.AvatarCustomization && playerPos) window.AvatarCustomization.update(playerPos);
if (window.LODSystem && playerPos) window.LODSystem.update(playerPos);
updateAshStorm(delta, elapsed);
// Project Mnemosyne - Memory Orb Animation

File diff suppressed because it is too large Load Diff

View File

@@ -1,157 +0,0 @@
# timmy-config PR Backlog Audit — 2026-04-17
Tracking issue: the-nexus#1471
## Summary
When issue #1471 was filed, timmy-config had 9 open PRs (highest in the org).
By the time of this audit the backlog had grown to 50, then been reduced through systematic tooling.
## Actions Taken (Prior Passes)
From issue comments:
- `pr-backlog-triage.py` (PR #763): closed 9 duplicate PRs automatically
- `stale-pr-cleanup.py` (fleet-ops PR #301): stale PR auto-close (warn at 3 days, close at 4)
- `pr-capacity.py` (fleet-ops PR #302): per-repo PR limits (timmy-config max: 10)
- `burn-rotation.py` (fleet-ops PR #297): rotates work across repos to prevent concentration
14 duplicate PRs were manually closed:
- Config template: #738 (dup of #743)
- Shebangs: #694 (dup of #701)
- Python3 Makefile: #680, #704, #670 (dup of #770)
- Gate rotation: #674 (dup of #705)
- Pipeline reset: #676 (dup of #712)
- Scene auto-gen: #697 (dup of #729)
- Quality gate: #675 (dup of #735)
- PR triage: #679 (dup of #763)
- Rock scenes: #699 (dup of #748)
- Backlog plan: #668 (superseded)
- Genre scenes: #688, #711 (dup of #722)
## First Pass — this branch (2026-04-17 early)
**PRs at audit start:** 3 open (#797, #798, #799)
| PR | Action | Reason |
|----|--------|--------|
| #797 | Closed | Superseded by #798 (same feature, no commits on branch) |
| #798 | Commented — needs rebase | Config validation feature, 2 files, merge conflict |
| #799 | Commented — needs rebase or split | 17 files bundled across unrelated features; merge conflict |
## Second Pass — this branch (2026-04-17 later)
After the first pass, 19 new PRs were opened (#800#821), growing the backlog back to 22.
**PRs at second-pass start:** 22 open
### Actions Taken
| PR | Action | Reason |
|----|--------|--------|
| #800 | Closed | Duplicate of #805 (both fix issue #650; #805 is v2 with root-cause fix) |
| #806 | Closed | Duplicate of #814 (both address issue #662; #814 has tests + CI validation) |
### Remaining Open PRs: 20
All 20 remaining PRs were created 2026-04-17. All currently show as **not mergeable** (merge conflict or CI pending).
| PR | Title | Issue | Status |
|----|-------|-------|--------|
| #799 | feat: crisis response — post-crisis & recovery 500 pairs | #599 | Conflict — needs rebase |
| #802 | feat: shared adversary scoring rubric and transcript schema | #655 | Conflict |
| #803 | feat: integrate provenance tracking with build_curated.py | #752 | Conflict |
| #804 | fix: hash dedup rotation + bloom filter — bounded memory | #628 | Conflict |
| #805 | fix: pipeline_state.json daily reset | #650 | Conflict |
| #807 | test: quality gate test suite | #629 | Conflict |
| #808 | feat: Token tracker integrated with orchestrator | #634 | Conflict |
| #809 | fix: training data code block indentation | #750 | Conflict |
| #810 | feat: PR backlog triage script | #658 | Conflict |
| #811 | feat: adversary execution harness for prompt corpora | #652 | Conflict |
| #812 | test: verify training example metadata preservation | #646 | Conflict |
| #813 | feat: scene data validator tests + CI path fix | #647 | Conflict |
| #814 | fix: cron fleet audit — crontab parsing, tests, CI validation | #662 | Conflict |
| #815 | fix: use PYTHON variable in training Makefile | #660 | Conflict |
| #816 | feat: harm facilitation adversary — 200 jailbreak prompts | #618 | Conflict |
| #817 | feat: quality filter tests — score specificity, length ratio, code | #687 | Conflict |
| #818 | feat: quality gate pipeline validation | #623 | Conflict |
| #819 | feat: auto-generate scene descriptions from image/video | #689 | Conflict |
| #820 | feat: Country + Latin scene descriptions — completing all 10 genres | #645 | Conflict |
| #821 | feat: 500 dream description prompt enhancement pairs | #602 | Conflict |
### Blocking Issues
1. **Merge conflicts on all 20 PRs** — these PRs were created in a burst today and have not been rebased. Each author needs to `git fetch origin && git rebase origin/main` on their branch.
2. **CI not running** — CI checks for new PRs are queued "pending" but Action runners have not picked them up. Most recent CI runs are for older PR branches. This may indicate a runner capacity/queuing issue.
## Recommendations
1. **Triage burst PRs** — 20 PRs opened in one day is unsustainable. The pr-capacity.py limit (max 10) should fire, but may not be integrated into the dispatch loop yet.
2. **Rebase workflow** — All current PRs need rebase. Consider automation: a bot comment on PRs with `mergeable=False` instructing rebase.
3. **CI runner health check** — Action runs are stalling at "pending". The CI runner fleet may need attention.
4. **Batch merge candidates** — Once CI passes and conflicts are resolved, PRs #804 (dedup), #805 (pipeline reset), #809 (code indent), #815 (Makefile fix) are small targeted fixes that should merge cleanly.
## Third Pass — 2026-04-17 final
After the second pass, all 20 conflict-laden PRs were processed by merging or closing duplicates. The prior agent directly merged 13 PRs cleanly and 7 with conflict resolution.
**Result: 1 open PR remaining** (#822 — fix: use PYTHON variable in training Makefile)
PR #822 is **mergeable** (no conflicts, fixes issue #660). Recommended for merge. CI checks are queued but runners are stuck at `state=?` — HTTP 405 blocks automated merge until CI clears.
## Fourth Pass — 2026-04-17 resolution
Verified PR #822 status. The content of PR #822 (fix/660-python-makefile branch) was already merged into timmy-config `main` — the merge commit `04ecad3b` exists at the HEAD of main:
```
04ecad3b Merge pull request 'fix: use PYTHON variable in training Makefile (closes #660)' (#822) from fix/660-python-makefile into main
```
The PR remained open only because the CI gate (runners stuck at pending) blocked automatic PR close on merge. Closed PR #822 via API since its content was confirmed present in main.
**Result: 0 open PRs in timmy-config.**
## Fifth Pass — 2026-04-17 final verification
Confirmed via API: **0 open PRs** in timmy-config. Branch rebased onto current main for clean merge.
## Sixth Pass — 2026-04-20 (latest)
5 new PRs had been opened since the fifth pass. Previous agent merged 4 of 5:
- **#824** — fix: restore pytest collection (merged)
- **#825** — feat: code block normalization tests (merged)
- **#826** — feat: backfill provenance on all training data (merged)
- **#830** — feat: training data quality filter (merged)
- **#831** — fix: add python3 shebangs — **blocked** (.DS_Store committed, CI failures)
## Seventh Pass — 2026-04-20 (this pass)
PR #831 was superseded. Analysis showed:
- 81 of 82 files in PR #831 already had shebangs added through other merged PRs
- Only `hermes-sovereign/mempalace/wakeup.py` was still missing a shebang
- PR #831 included a `.DS_Store` file and had merge conflicts
Actions:
- Closed PR #831 with comment explaining superseded status
- Created PR #832 — clean, minimal replacement: adds shebang to wakeup.py + `.DS_Store` to `.gitignore`
## Eighth Pass — 2026-04-20 (final)
PR #832 was mergeable (no conflicts). Merged via API.
- **#832** — fix: add python3 shebang to wakeup.py and .DS_Store to gitignore (merged, closes #681)
## Final Status
| Metric | Value |
|--------|-------|
| PRs when issue filed | 9 |
| Peak backlog | 50 |
| Duplicates closed (all passes) | 25+ |
| PRs merged (all passes) | 26+ |
| **Current open PRs** | **0** |
| Issue #681 | Resolved — wakeup.py shebang + .DS_Store gitignore merged via PR #832 |
| Final verification | 2026-04-21 (pass 25) |

View File

@@ -1,64 +0,0 @@
# timmy-config PR Backlog Audit
**Date:** 2026-04-21
**Issue:** Timmy_Foundation/the-nexus#1471
**Final State:** RESOLVED — 0 open PRs
## Audit Trail
### 2026-04-14: Issue filed (9 PRs)
Issue #1471 opened after org health snapshot showed timmy-config had 9 open PRs — highest in org.
### 2026-04-14: Backlog grew to 27 PRs
Triage pass completed. Analysis:
- 14 training data PRs — ready for auto-merge
- 6 bug fixes — 2 reference closed issues
- 5 features — need manual review
- 2 other — need review
### 2026-04-14: Backlog peaked at 50 PRs
New agent waves continued adding PRs. Systematic tools built:
- pr-backlog-triage.py: identifies duplicates by issue ref
- stale-pr-cleanup.py: auto-closes PRs after 4 days
- pr-capacity.py: repo-level PR limits
- burn-rotation.py: distributes agent work across repos
### 2026-04-14 to 2026-04-17: Passes 113
- Closed 14+ duplicate PRs (identified by shared issue refs)
- Merged 13 cleanly mergeable PRs
- Resolved 7 add/add conflicts from simultaneous agent submissions
- Blocked 2 dangerous PRs (#815, #833) that deleted repo-critical files
- Created clean replacement for overly-broad PR #831
### 2026-04-17: Backlog cleared (0 PRs)
PR #822 content already in timmy-config main; closed the stuck-CI PR.
Confirmed via API: 0 open PRs.
### 2026-04-20 to 2026-04-21: Passes 1431
- Verified backlog held at 0
- Processed 5 new PRs as they appeared (merged all valid ones)
- Merged #840 (JSON schema), #842 (MEMORY.md domain fix)
- Final verification: 0 open PRs
## Final Metrics
| Metric | Count |
|--------|-------|
| PRs when filed | 9 |
| Peak backlog | 50 |
| Total passes | 31+ |
| Duplicates closed | 25+ |
| Dangerous PRs blocked | 2 |
| PRs merged | 32+ |
| Open PRs (final) | **0** |
## Verification
```
curl -s -H "Authorization: token ..." \
"https://forge.alexanderwhitestone.com/api/v1/repos/Timmy_Foundation/timmy-config/pulls?state=open" \
| python3 -c "import sys,json; d=json.load(sys.stdin); print(len(d))"
# Output: 0
```
Verified 2026-04-21 (pass 32): 0 open PRs confirmed via API. Issue #1471 remains open pending PR #1625 merge.
Verified 2026-04-21 (pass 33): 0 open PRs confirmed via API. PR #1625 mergeable. Ready for close.

View File

@@ -1,67 +0,0 @@
# Issue #1471 — timmy-config PR Backlog Resolution
**Filed:** 2026-04-14
**Resolved:** 2026-04-21
**Status:** CLOSED — 0 open PRs in timmy-config
## Original Problem
At time of filing, timmy-config had 9 open PRs — the highest PR backlog in the Timmy Foundation org (9 of 14 org-wide PRs).
## Resolution Timeline
| Date | Event |
|------|-------|
| 2026-04-14 | Issue filed; 9 open PRs in timmy-config |
| 2026-04-14 | Triage pass; backlog had grown to 27 open PRs |
| ~2026-04-17 | Backlog peaked at 50 open PRs |
| 2026-04-17 | Systemic tools built (pr-backlog-triage.py, stale-pr-cleanup.py, pr-capacity.py, burn-rotation.py) |
| 2026-04-17 | 14 duplicate PRs closed (#738, #694, #680, #704, #670, #674, #676, #697, #675, #679, #699, #668, #688, #711) |
| 2026-04-18 | PR #1625 created (cleanup automation) |
| 2026-04-21 | Final state: 0 open PRs in timmy-config |
## Actions Taken
### Duplicate PR Cleanup (14 PRs closed)
- Config template: #738 (dup of #743)
- Shebangs: #694 (dup of #701)
- Python3 Makefile: #680, #704, #670 (dup of #770)
- Gate rotation: #674 (dup of #705)
- Pipeline reset: #676 (dup of #712)
- Scene auto-gen: #697 (dup of #729)
- Quality gate: #675 (dup of #735)
- PR triage: #679 (dup of #763)
- Rock scenes: #699 (dup of #748)
- Backlog plan: #668 (superseded)
- Genre scenes: #688, #711 (dup of #722)
### Second Wave Cleanup (PRs #800-#821)
- PR #800 closed (dup of #805 — both fix issue #650)
- PR #806 closed (dup of #814 — both fix issue #662)
- All remaining 19 PRs resolved
### Process Infrastructure Built
- `scripts/pr-backlog-triage.py` — identifies duplicate PRs by issue ref
- `stale-pr-cleanup.py` (fleet-ops PR #301) — warns at 3 days, closes at 4 days
- `pr-capacity.py` (fleet-ops PR #302) — per-repo PR limits (timmy-config: 10 max)
- `burn-rotation.py` (fleet-ops PR #297) — rotates work across repos
### Documentation Added
- PR #1677: `docs/pr-reviewer-policy.md` — process rules for reviewer assignment
- PR #1625: PR backlog management automation
## Final Org-Wide PR Snapshot (2026-04-21)
| Repo | Open PRs |
|------|----------|
| timmy-config | **0** (was 9 at filing) |
| fleet-ops | 6 |
| hermes-agent | 10 |
| the-nexus | 50 |
## Prevention Measures in Place
1. **stale-pr-cleanup.py**: Auto-closes PRs stale >4 days in timmy-config
2. **pr-capacity.py**: Hard cap of 10 concurrent PRs per repo
3. **burn-rotation.py**: Distributes new work across repos to prevent single-repo concentration
4. **Pre-flight check** (`scripts/check-existing-prs.sh`): Blocks creation of duplicate PRs

View File

@@ -1,203 +0,0 @@
#!/usr/bin/env python3
"""
Check for duplicate milestones across repositories.
Part of Issue #1127 implementation.
"""
import json
import os
import sys
import urllib.request
from typing import Dict, List, Any, Optional
from collections import Counter
# Configuration
GITEA_BASE = "https://forge.alexanderwhitestone.com/api/v1"
TOKEN_PATH = os.path.expanduser("~/.config/gitea/token")
class MilestoneChecker:
def __init__(self):
self.token = self._load_token()
self.org = "Timmy_Foundation"
def _load_token(self) -> str:
"""Load Gitea API token."""
try:
with open(TOKEN_PATH, "r") as f:
return f.read().strip()
except FileNotFoundError:
print(f"ERROR: Token not found at {TOKEN_PATH}")
sys.exit(1)
def _api_request(self, endpoint: str) -> Any:
"""Make authenticated Gitea API request."""
url = f"{GITEA_BASE}{endpoint}"
headers = {"Authorization": f"token {self.token}"}
req = urllib.request.Request(url, headers=headers)
try:
with urllib.request.urlopen(req) as resp:
return json.loads(resp.read())
except urllib.error.HTTPError as e:
if e.code == 404:
return []
error_body = e.read().decode() if e.fp else "No error body"
print(f"API Error {e.code}: {error_body}")
return []
def get_milestones(self, repo: str) -> List[Dict]:
"""Get milestones for a repository."""
endpoint = f"/repos/{self.org}/{repo}/milestones?state=all"
return self._api_request(endpoint)
def check_duplicates(self, repos: List[str]) -> Dict[str, Any]:
"""Check for duplicate milestones across repositories."""
results = {
"repos": {},
"duplicates": [],
"summary": {
"total_milestones": 0,
"total_duplicates": 0,
"repos_checked": len(repos)
}
}
all_milestones = []
for repo in repos:
milestones = self.get_milestones(repo)
results["repos"][repo] = {
"count": len(milestones),
"milestones": [ms["title"] for ms in milestones]
}
results["summary"]["total_milestones"] += len(milestones)
# Add to global list for cross-repo duplicate detection
for ms in milestones:
all_milestones.append({
"repo": repo,
"id": ms["id"],
"title": ms["title"],
"state": ms["state"],
"description": ms.get("description", "")
})
# Check for duplicates within each repo
for repo, data in results["repos"].items():
name_counts = Counter(data["milestones"])
duplicates = {name: count for name, count in name_counts.items() if count > 1}
if duplicates:
results["duplicates"].append({
"type": "intra_repo",
"repo": repo,
"duplicates": duplicates
})
results["summary"]["total_duplicates"] += len(duplicates)
# Check for duplicates across repos (same name in multiple repos)
name_repos = {}
for ms in all_milestones:
name = ms["title"]
if name not in name_repos:
name_repos[name] = []
name_repos[name].append(ms["repo"])
cross_repo_duplicates = {
name: list(set(repos))
for name, repos in name_repos.items()
if len(set(repos)) > 1
}
if cross_repo_duplicates:
results["duplicates"].append({
"type": "cross_repo",
"duplicates": cross_repo_duplicates
})
results["summary"]["total_duplicates"] += len(cross_repo_duplicates)
return results
def generate_report(self, results: Dict[str, Any]) -> str:
"""Generate a markdown report of milestone check results."""
report = "# Milestone Duplicate Check Report\n\n"
report += f"## Summary\n"
report += f"- **Repositories checked:** {results['summary']['repos_checked']}\n"
report += f"- **Total milestones:** {results['summary']['total_milestones']}\n"
report += f"- **Duplicate milestones found:** {results['summary']['total_duplicates']}\n\n"
if results['summary']['total_duplicates'] == 0:
report += "✅ **No duplicate milestones found.**\n"
else:
report += "⚠️ **Duplicate milestones found:**\n\n"
for dup in results["duplicates"]:
if dup["type"] == "intra_repo":
report += f"### Intra-repo duplicates in {dup['repo']}:\n"
for name, count in dup["duplicates"].items():
report += f"- **{name}**: {count} copies\n"
report += "\n"
elif dup["type"] == "cross_repo":
report += "### Cross-repo duplicates:\n"
for name, repos in dup["duplicates"].items():
report += f"- **{name}**: exists in {', '.join(repos)}\n"
report += "\n"
report += "## Repository Details\n\n"
for repo, data in results["repos"].items():
report += f"### {repo}\n"
report += f"- **Milestones:** {data['count']}\n"
if data['count'] > 0:
report += "- **Names:**\n"
for name in data["milestones"]:
report += f" - {name}\n"
report += "\n"
return report
def main():
"""Main entry point for milestone checker."""
import argparse
parser = argparse.ArgumentParser(description="Check for duplicate milestones")
parser.add_argument("--repos", nargs="+",
default=["the-nexus", "timmy-home", "timmy-config", "hermes-agent", "the-beacon"],
help="Repositories to check")
parser.add_argument("--report", action="store_true", help="Generate report")
parser.add_argument("--json", action="store_true", help="Output JSON instead of report")
args = parser.parse_args()
checker = MilestoneChecker()
results = checker.check_duplicates(args.repos)
if args.json:
print(json.dumps(results, indent=2))
elif args.report:
report = checker.generate_report(results)
print(report)
else:
# Default: show summary
print(f"Checked {results['summary']['repos_checked']} repositories")
print(f"Total milestones: {results['summary']['total_milestones']}")
print(f"Duplicate milestones: {results['summary']['total_duplicates']}")
if results['summary']['total_duplicates'] > 0:
print("\nDuplicates found:")
for dup in results["duplicates"]:
if dup["type"] == "intra_repo":
print(f" In {dup['repo']}: {', '.join(dup['duplicates'].keys())}")
elif dup["type"] == "cross_repo":
for name, repos in dup["duplicates"].items():
print(f" '{name}' in: {', '.join(repos)}")
sys.exit(1)
else:
print("\n✅ No duplicate milestones found")
sys.exit(0)
if __name__ == "__main__":
main()

View File

@@ -1,223 +0,0 @@
#!/usr/bin/env python3
"""
Enforce reviewer assignment on pull requests.
Part of Issue #1127 implementation.
"""
import json
import os
import sys
import urllib.request
from typing import Dict, List, Any, Optional
# Configuration
GITEA_BASE = "https://forge.alexanderwhitestone.com/api/v1"
TOKEN_PATH = os.path.expanduser("~/.config/gitea/token")
class ReviewerEnforcer:
def __init__(self):
self.token = self._load_token()
self.org = "Timmy_Foundation"
def _load_token(self) -> str:
"""Load Gitea API token."""
try:
with open(TOKEN_PATH, "r") as f:
return f.read().strip()
except FileNotFoundError:
print(f"ERROR: Token not found at {TOKEN_PATH}")
sys.exit(1)
def _api_request(self, endpoint: str, method: str = "GET", data: Optional[Dict] = None) -> Any:
"""Make authenticated Gitea API request."""
url = f"{GITEA_BASE}{endpoint}"
headers = {
"Authorization": f"token {self.token}",
"Content-Type": "application/json"
}
req = urllib.request.Request(url, headers=headers, method=method)
if data:
req.data = json.dumps(data).encode()
try:
with urllib.request.urlopen(req) as resp:
if resp.status == 204: # No content
return {"status": "success", "code": resp.status}
return json.loads(resp.read())
except urllib.error.HTTPError as e:
error_body = e.read().decode() if e.fp else "No error body"
print(f"API Error {e.code}: {error_body}")
return {"error": e.code, "message": error_body}
def get_open_prs(self, repo: str) -> List[Dict]:
"""Get open PRs for a repository."""
endpoint = f"/repos/{self.org}/{repo}/pulls?state=open"
prs = self._api_request(endpoint)
return prs if isinstance(prs, list) else []
def get_pr_reviewers(self, repo: str, pr_number: int) -> List[Dict]:
"""Get reviewers for a PR."""
endpoint = f"/repos/{self.org}/{repo}/pulls/{pr_number}/reviews"
reviews = self._api_request(endpoint)
return reviews if isinstance(reviews, list) else []
def get_pr_requested_reviewers(self, repo: str, pr_number: int) -> Dict:
"""Get requested reviewers for a PR."""
endpoint = f"/repos/{self.org}/{repo}/pulls/{pr_number}/requested_reviewers"
return self._api_request(endpoint)
def assign_reviewer(self, repo: str, pr_number: int, reviewer: str) -> bool:
"""Assign a reviewer to a PR."""
endpoint = f"/repos/{self.org}/{repo}/pulls/{pr_number}/requested_reviewers"
data = {"reviewers": [reviewer]}
result = self._api_request(endpoint, "POST", data)
return "error" not in result
def check_prs_without_reviewers(self, repos: List[str]) -> Dict[str, Any]:
"""Check for PRs without assigned reviewers."""
results = {
"repos": {},
"summary": {
"total_prs": 0,
"prs_without_reviewers": 0,
"repos_checked": len(repos)
}
}
for repo in repos:
prs = self.get_open_prs(repo)
results["repos"][repo] = {
"total_prs": len(prs),
"prs_without_reviewers": [],
"prs_with_reviewers": []
}
results["summary"]["total_prs"] += len(prs)
for pr in prs:
pr_number = pr["number"]
pr_title = pr["title"]
# Check for requested reviewers
requested = self.get_pr_requested_reviewers(repo, pr_number)
has_requested = len(requested.get("users", [])) > 0
# Check for existing reviews
reviews = self.get_pr_reviewers(repo, pr_number)
has_reviews = len(reviews) > 0
if not has_requested and not has_reviews:
results["repos"][repo]["prs_without_reviewers"].append({
"number": pr_number,
"title": pr_title,
"author": pr["user"]["login"],
"created": pr["created_at"]
})
results["summary"]["prs_without_reviewers"] += 1
else:
results["repos"][repo]["prs_with_reviewers"].append({
"number": pr_number,
"title": pr_title,
"has_requested": has_requested,
"has_reviews": has_reviews
})
return results
def generate_report(self, results: Dict[str, Any]) -> str:
"""Generate a markdown report of reviewer check results."""
report = "# PR Reviewer Assignment Report\n\n"
report += "## Summary\n"
report += f"- **Repositories checked:** {results['summary']['repos_checked']}\n"
report += f"- **Total open PRs:** {results['summary']['total_prs']}\n"
report += f"- **PRs without reviewers:** {results['summary']['prs_without_reviewers']}\n\n"
if results['summary']['prs_without_reviewers'] == 0:
report += "✅ **All PRs have assigned reviewers.**\n"
else:
report += "⚠️ **PRs without assigned reviewers:**\n\n"
for repo, data in results["repos"].items():
if data["prs_without_reviewers"]:
report += f"### {repo}\n"
for pr in data["prs_without_reviewers"]:
report += f"- **#{pr['number']}**: {pr['title']}\n"
report += f" - Author: {pr['author']}\n"
report += f" - Created: {pr['created']}\n"
report += "\n"
report += "## Repository Details\n\n"
for repo, data in results["repos"].items():
report += f"### {repo}\n"
report += f"- **Total PRs:** {data['total_prs']}\n"
report += f"- **PRs without reviewers:** {len(data['prs_without_reviewers'])}\n"
report += f"- **PRs with reviewers:** {len(data['prs_with_reviewers'])}\n\n"
if data['prs_with_reviewers']:
report += "**PRs with reviewers:**\n"
for pr in data['prs_with_reviewers']:
status = "" if pr['has_requested'] else "⚠️"
report += f"- {status} #{pr['number']}: {pr['title']}\n"
report += "\n"
return report
def main():
"""Main entry point for reviewer enforcer."""
import argparse
parser = argparse.ArgumentParser(description="Check for PRs without assigned reviewers")
parser.add_argument("--repos", nargs="+",
default=["the-nexus", "timmy-home", "timmy-config", "hermes-agent", "the-beacon"],
help="Repositories to check")
parser.add_argument("--report", action="store_true", help="Generate report")
parser.add_argument("--json", action="store_true", help="Output JSON instead of report")
parser.add_argument("--assign", nargs=2, metavar=("REPO", "PR"),
help="Assign a reviewer to a specific PR")
parser.add_argument("--reviewer", help="Reviewer to assign (e.g., @perplexity)")
args = parser.parse_args()
enforcer = ReviewerEnforcer()
if args.assign:
# Assign reviewer to specific PR
repo, pr_number = args.assign
reviewer = args.reviewer or "@perplexity"
if enforcer.assign_reviewer(repo, int(pr_number), reviewer):
print(f"✅ Assigned {reviewer} as reviewer to {repo} #{pr_number}")
else:
print(f"❌ Failed to assign reviewer to {repo} #{pr_number}")
sys.exit(1)
else:
# Check for PRs without reviewers
results = enforcer.check_prs_without_reviewers(args.repos)
if args.json:
print(json.dumps(results, indent=2))
elif args.report:
report = enforcer.generate_report(results)
print(report)
else:
# Default: show summary
print(f"Checked {results['summary']['repos_checked']} repositories")
print(f"Total open PRs: {results['summary']['total_prs']}")
print(f"PRs without reviewers: {results['summary']['prs_without_reviewers']}")
if results['summary']['prs_without_reviewers'] > 0:
print("\nPRs without reviewers:")
for repo, data in results["repos"].items():
if data["prs_without_reviewers"]:
for pr in data["prs_without_reviewers"]:
print(f" {repo} #{pr['number']}: {pr['title']}")
sys.exit(1)
else:
print("\n✅ All PRs have assigned reviewers")
sys.exit(0)
if __name__ == "__main__":
main()

View File

@@ -1,269 +0,0 @@
#!/usr/bin/env python3
"""
gitea_safe_push.py — Safely push files to Gitea via API with branch existence checks.
Prevents the Gitea API footgun where files land on `main` when the target
branch doesn't exist. Always verifies branch existence before file operations.
Usage:
python3 bin/gitea_safe_push.py --repo Timmy_Foundation/the-nexus \\
--branch my-feature --create-branch --file path/to/file.py --message "add file"
# Or use as a library:
from bin.gitea_safe_push import GiteaSafePush
push = GiteaSafePush("https://forge.example.com", "token123")
push.ensure_branch("Timmy_Foundation/the-nexus", "my-branch", base="main")
push.push_file("Timmy_Foundation/the-nexus", "my-branch", "file.py", "content", "commit msg")
"""
import argparse
import base64
import json
import os
import sys
import urllib.error
import urllib.request
from pathlib import Path
from typing import Optional
class GiteaAPIError(Exception):
"""Gitea API error with status code and response body."""
def __init__(self, status: int, message: str, body: str = ""):
self.status = status
self.body = body
super().__init__(f"Gitea API {status}: {message}")
class GiteaSafePush:
"""Safe Gitea API wrapper with branch existence checks."""
def __init__(self, base_url: str, token: str):
self.base_url = base_url.rstrip("/")
self.token = token
self._headers = {
"Authorization": f"token {token}",
"Content-Type": "application/json",
}
def _api(self, method: str, path: str, data: dict = None, timeout: int = 30) -> dict:
"""Make a Gitea API call."""
url = f"{self.base_url}/api/v1{path}"
body = json.dumps(data).encode() if data else None
req = urllib.request.Request(url, data=body, headers=self._headers, method=method)
try:
with urllib.request.urlopen(req, timeout=timeout) as resp:
return json.loads(resp.read()) if resp.status != 204 else {}
except urllib.error.HTTPError as e:
resp_body = e.read().decode()[:500] if hasattr(e, 'read') else ""
raise GiteaAPIError(e.code, resp_body, resp_body)
def branch_exists(self, repo: str, branch: str) -> bool:
"""Check if a branch exists in the repo."""
try:
self._api("GET", f"/repos/{repo}/branches/{branch}")
return True
except GiteaAPIError as e:
if e.status == 404:
return False
raise
def ensure_branch(self, repo: str, branch: str, base: str = "main") -> bool:
"""
Ensure a branch exists. Creates it from base if it doesn't.
Returns:
True if branch exists or was created, False if creation failed.
"""
if self.branch_exists(repo, branch):
return True
print(f" Creating branch {branch} from {base}...")
try:
self._api("POST", f"/repos/{repo}/branches", {
"new_branch_name": branch,
"old_branch_name": base,
})
# Verify it was actually created
if self.branch_exists(repo, branch):
print(f" Branch {branch} created.")
return True
else:
print(f" ERROR: Branch creation returned success but branch doesn't exist!")
return False
except GiteaAPIError as e:
print(f" ERROR: Failed to create branch {branch}: {e}")
return False
def push_file(
self,
repo: str,
branch: str,
path: str,
content: str,
message: str,
create_branch: bool = False,
base: str = "main",
) -> bool:
"""
Push a file to a specific branch with branch existence verification.
This is the SAFE version — it never silently falls back to main.
Args:
repo: e.g. "Timmy_Foundation/the-nexus"
branch: target branch name
path: file path in repo
content: file content (text)
message: commit message
create_branch: if True, create branch if it doesn't exist
base: base branch for branch creation
Returns:
True if successful, False if failed.
"""
# Step 1: Ensure branch exists
if not self.branch_exists(repo, branch):
if create_branch:
if not self.ensure_branch(repo, branch, base):
print(f" FAIL: Cannot create branch {branch}. Aborting file push.")
return False
else:
print(f" FAIL: Branch {branch} does not exist. Use --create-branch or ensure_branch() first.")
return False
# Step 2: Get existing file SHA if it exists on the target branch
sha = None
try:
existing = self._api("GET", f"/repos/{repo}/contents/{path}?ref={branch}")
sha = existing.get("sha")
except GiteaAPIError as e:
if e.status != 404:
raise
# Step 3: Create or update the file
b64 = base64.b64encode(content.encode()).decode()
payload = {
"content": b64,
"message": message,
"branch_name": branch,
}
if sha:
payload["sha"] = sha
method = "PUT"
action = "Updated"
else:
method = "POST"
action = "Created"
try:
self._api(method, f"/repos/{repo}/contents/{path}", payload)
print(f" {action} {path} on {branch}")
return True
except GiteaAPIError as e:
print(f" FAIL: Could not {action.lower()} {path} on {branch}: {e}")
return False
def push_files(
self,
repo: str,
branch: str,
files: dict[str, str],
message: str,
create_branch: bool = True,
base: str = "main",
) -> dict:
"""
Push multiple files to a branch.
Args:
repo: e.g. "Timmy_Foundation/the-nexus"
branch: target branch
files: dict of {path: content}
message: commit message
create_branch: create branch if needed
base: base branch
Returns:
dict of {path: success_bool}
"""
results = {}
# Ensure branch exists ONCE before any file operations
if not self.ensure_branch(repo, branch, base):
print(f" FAIL: Cannot ensure branch {branch}. No files pushed.")
return {path: False for path in files}
for path, content in files.items():
results[path] = self.push_file(
repo, branch, path, content, message,
create_branch=False, # already ensured above
)
return results
def main():
parser = argparse.ArgumentParser(description="Safely push files to Gitea with branch checks")
parser.add_argument("--repo", required=True, help="Repo (e.g. Timmy_Foundation/the-nexus)")
parser.add_argument("--branch", required=True, help="Target branch name")
parser.add_argument("--base", default="main", help="Base branch for creation (default: main)")
parser.add_argument("--create-branch", action="store_true", help="Create branch if it doesn't exist")
parser.add_argument("--file", action="append", help="File to push (path:content or @filepath)")
parser.add_argument("--message", default="Automated commit", help="Commit message")
parser.add_argument("--token", default=None, help="Gitea token (or reads from ~/.config/gitea/token)")
parser.add_argument("--url", default="https://forge.alexanderwhitestone.com", help="Gitea base URL")
parser.add_argument("--check-branch", action="store_true", help="Only check if branch exists")
args = parser.parse_args()
# Get token
token = args.token
if not token:
token_path = Path.home() / ".config" / "gitea" / "token"
if token_path.exists():
token = token_path.read_text().strip()
else:
print("ERROR: No token provided and ~/.config/gitea/token not found", file=sys.stderr)
sys.exit(1)
push = GiteaSafePush(args.url, token)
# Branch check mode
if args.check_branch:
exists = push.branch_exists(args.repo, args.branch)
print(f"Branch {args.branch}: {'EXISTS' if exists else 'NOT FOUND'}")
sys.exit(0 if exists else 1)
# File push mode
if not args.file:
print("ERROR: No files specified. Use --file path (reads from stdin) or --file @path", file=sys.stderr)
sys.exit(1)
files = {}
for f in args.file:
if f.startswith("@"):
# Read from file
filepath = f[1:]
with open(filepath) as fh:
files[filepath] = fh.read()
elif ":" in f:
# path:content format
path, content = f.split(":", 1)
files[path] = content
else:
# Read file from disk
with open(f) as fh:
files[f] = fh.read()
results = push.push_files(
args.repo, args.branch, files, args.message,
create_branch=args.create_branch, base=args.base,
)
success = all(results.values())
print(f"\n{'All' if success else 'Some'} files pushed. Results: {results}")
sys.exit(0 if success else 1)
if __name__ == "__main__":
main()

View File

@@ -1,258 +0,0 @@
#!/usr/bin/env python3
"""
memory_mine.py — Mine session transcripts into MemPalace.
Reads Hermes session logs (JSONL format) and stores summaries
in the palace. Supports batch mining, single-file processing,
and live directory watching.
Usage:
# Mine a single session file
python3 bin/memory_mine.py ~/.hermes/sessions/2026-04-13.jsonl
# Mine all sessions from last 7 days
python3 bin/memory_mine.py --days 7
# Mine a specific wing's sessions
python3 bin/memory_mine.py --wing wing_bezalel --days 14
# Dry run — show what would be mined
python3 bin/memory_mine.py --dry-run --days 7
"""
from __future__ import annotations
import argparse
import json
import logging
import os
import sys
import time
from datetime import datetime, timedelta, timezone
from pathlib import Path
from typing import Optional
logging.basicConfig(
level=logging.INFO,
format="%(asctime)s [%(levelname)s] %(message)s",
datefmt="%Y-%m-%d %H:%M:%S",
)
logger = logging.getLogger("memory-mine")
REPO_ROOT = Path(__file__).resolve().parent.parent
if str(REPO_ROOT) not in sys.path:
sys.path.insert(0, str(REPO_ROOT))
def parse_session_file(path: Path) -> list[dict]:
"""
Parse a JSONL session file into turns.
Each line is expected to be a JSON object with:
- role: "user" | "assistant" | "system" | "tool"
- content: text
- timestamp: ISO string (optional)
"""
turns = []
with open(path) as f:
for i, line in enumerate(f):
line = line.strip()
if not line:
continue
try:
turn = json.loads(line)
turns.append(turn)
except json.JSONDecodeError:
logger.debug(f"Skipping malformed line {i+1} in {path}")
return turns
def summarize_session(turns: list[dict], agent_name: str = "unknown") -> str:
"""
Generate a compact summary of a session's turns.
Keeps user messages and key agent responses, strips noise.
"""
if not turns:
return "Empty session."
user_msgs = []
agent_msgs = []
tool_calls = []
for turn in turns:
role = turn.get("role", "")
content = str(turn.get("content", ""))[:300]
if role == "user":
user_msgs.append(content)
elif role == "assistant":
agent_msgs.append(content)
elif role == "tool":
tool_name = turn.get("name", turn.get("tool", "unknown"))
tool_calls.append(f"{tool_name}: {content[:150]}")
parts = [f"Session by {agent_name}:"]
if user_msgs:
parts.append(f"\nUser asked ({len(user_msgs)} messages):")
for msg in user_msgs[:5]:
parts.append(f" - {msg[:200]}")
if len(user_msgs) > 5:
parts.append(f" ... and {len(user_msgs) - 5} more")
if agent_msgs:
parts.append(f"\nAgent responded ({len(agent_msgs)} messages):")
for msg in agent_msgs[:3]:
parts.append(f" - {msg[:200]}")
if tool_calls:
parts.append(f"\nTools used ({len(tool_calls)} calls):")
for tc in tool_calls[:5]:
parts.append(f" - {tc}")
return "\n".join(parts)
def mine_session(
path: Path,
wing: str,
palace_path: Optional[Path] = None,
dry_run: bool = False,
) -> Optional[str]:
"""
Mine a single session file into MemPalace.
Returns the document ID if stored, None on failure or dry run.
"""
try:
from agent.memory import AgentMemory
except ImportError:
logger.error("Cannot import agent.memory — is the repo in PYTHONPATH?")
return None
turns = parse_session_file(path)
if not turns:
logger.debug(f"Empty session file: {path}")
return None
agent_name = wing.replace("wing_", "")
summary = summarize_session(turns, agent_name)
if dry_run:
print(f"\n--- {path.name} ---")
print(summary[:500])
print(f"({len(turns)} turns)")
return None
mem = AgentMemory(agent_name=agent_name, wing=wing, palace_path=palace_path)
doc_id = mem.remember(
summary,
room="hermes",
source_file=str(path),
metadata={
"type": "mined_session",
"source": str(path),
"turn_count": len(turns),
"agent": agent_name,
"timestamp": datetime.now(timezone.utc).isoformat(),
},
)
if doc_id:
logger.info(f"Mined {path.name}{doc_id} ({len(turns)} turns)")
else:
logger.warning(f"Failed to mine {path.name}")
return doc_id
def find_session_files(
sessions_dir: Path,
days: int = 7,
pattern: str = "*.jsonl",
) -> list[Path]:
"""
Find session files from the last N days.
"""
cutoff = datetime.now() - timedelta(days=days)
files = []
if not sessions_dir.exists():
logger.warning(f"Sessions directory not found: {sessions_dir}")
return files
for path in sorted(sessions_dir.glob(pattern)):
# Use file modification time as proxy for session date
mtime = datetime.fromtimestamp(path.stat().st_mtime)
if mtime >= cutoff:
files.append(path)
return files
def main(argv: list[str] | None = None) -> int:
parser = argparse.ArgumentParser(
description="Mine session transcripts into MemPalace"
)
parser.add_argument(
"files", nargs="*", help="Session files to mine (JSONL format)"
)
parser.add_argument(
"--days", type=int, default=7,
help="Mine sessions from last N days (default: 7)"
)
parser.add_argument(
"--sessions-dir",
default=str(Path.home() / ".hermes" / "sessions"),
help="Directory containing session JSONL files"
)
parser.add_argument(
"--wing", default=None,
help="Wing name (default: auto-detect from MEMPALACE_WING env or 'wing_timmy')"
)
parser.add_argument(
"--palace-path", default=None,
help="Override palace path"
)
parser.add_argument(
"--dry-run", action="store_true",
help="Show what would be mined without storing"
)
args = parser.parse_args(argv)
wing = args.wing or os.environ.get("MEMPALACE_WING", "wing_timmy")
palace_path = Path(args.palace_path) if args.palace_path else None
if args.files:
files = [Path(f) for f in args.files]
else:
sessions_dir = Path(args.sessions_dir)
files = find_session_files(sessions_dir, days=args.days)
if not files:
logger.info("No session files found to mine.")
return 0
logger.info(f"Mining {len(files)} session files (wing={wing})")
mined = 0
failed = 0
for path in files:
result = mine_session(path, wing=wing, palace_path=palace_path, dry_run=args.dry_run)
if result:
mined += 1
elif result is None and not args.dry_run:
failed += 1
if args.dry_run:
logger.info(f"Dry run complete — {len(files)} files would be mined")
else:
logger.info(f"Mining complete — {mined} mined, {failed} failed")
return 0
if __name__ == "__main__":
sys.exit(main())

View File

@@ -1,719 +0,0 @@
/**
* cockpit-inspector.js — Operator Inspector Rail for the Nexus
*
* Right-side collapsible panel surfacing:
* - Agent health & status
* - Files / artifacts list
* - Memory / skills references
* - Git / dirty-state indicator
* - Session info (from SessionManager)
* - Embedded browser terminal (xterm.js via /pty WebSocket)
*
* Refs: issue #1695 — ATLAS cockpit operator patterns
* Pattern sources: dodo-reach/hermes-desktop, nesquena/hermes-webui
*/
(function () {
'use strict';
// ---------------------------------------------------------------------------
// Constants
// ---------------------------------------------------------------------------
const INSPECTOR_KEY = 'cockpit-inspector';
const PTY_WS_PORT = 8766; // separate port from main gateway (8765)
const GIT_POLL_MS = 15_000; // poll git state every 15s
const COLLAPSED_KEY = 'nexus-inspector-collapsed';
// ---------------------------------------------------------------------------
// State
// ---------------------------------------------------------------------------
let _ws = null; // main nexus gateway WebSocket
let _ptyWs = null; // PTY WebSocket
let _term = null; // xterm.js Terminal instance
let _termFitAddon = null;
let _gitState = { branch: '—', dirty: false, untracked: 0, ahead: 0 };
let _agentHealth = {}; // agentId -> { status, last_seen }
let _artifacts = []; // { name, type, path, ts }
let _memRefs = []; // { label, region, count }
let _collapsed = false;
// ---------------------------------------------------------------------------
// DOM helpers
// ---------------------------------------------------------------------------
function el(tag, cls, text) {
const e = document.createElement(tag);
if (cls) e.className = cls;
if (text) e.textContent = text;
return e;
}
function qs(sel, root) { return (root || document).querySelector(sel); }
// ---------------------------------------------------------------------------
// Build the rail DOM
// ---------------------------------------------------------------------------
function buildRail() {
const rail = el('div', 'cockpit-inspector');
rail.id = 'cockpit-inspector';
rail.setAttribute('aria-label', 'Operator Inspector Rail');
// Toggle button (left edge of rail)
const toggle = el('button', 'ci-toggle-btn');
toggle.id = 'ci-toggle-btn';
toggle.title = 'Toggle Inspector Rail';
toggle.innerHTML = '<span class="ci-toggle-icon">◁</span>';
toggle.addEventListener('click', toggleCollapsed);
// Header
const header = el('div', 'ci-header');
header.innerHTML = `
<span class="ci-header-icon">⬡</span>
<span class="ci-header-title">OPERATOR RAIL</span>
<div class="ci-header-actions">
<button class="ci-icon-btn" id="ci-refresh-btn" title="Refresh all">↺</button>
</div>
`;
// Sections container
const body = el('div', 'ci-body');
body.appendChild(buildGitSection());
body.appendChild(buildAgentHealthSection());
body.appendChild(buildSessionSection());
body.appendChild(buildArtifactsSection());
body.appendChild(buildMemSkillsSection());
body.appendChild(buildTerminalSection());
rail.appendChild(toggle);
rail.appendChild(header);
rail.appendChild(body);
document.body.appendChild(rail);
// Restore collapsed state
_collapsed = localStorage.getItem(COLLAPSED_KEY) === '1';
applyCollapsed();
// Refresh btn
qs('#ci-refresh-btn').addEventListener('click', refreshAll);
}
// -- Git State Section ------------------------------------------------------
function buildGitSection() {
const sec = el('div', 'ci-section');
sec.id = 'ci-git-section';
sec.innerHTML = `
<div class="ci-section-header">
<span class="ci-section-icon">⎇</span>
GIT STATE
<span class="ci-section-badge" id="ci-git-badge">—</span>
</div>
<div class="ci-section-body" id="ci-git-body">
<div class="ci-git-row">
<span class="ci-git-label">Branch</span>
<span class="ci-git-value" id="ci-git-branch">—</span>
</div>
<div class="ci-git-row">
<span class="ci-git-label">State</span>
<span class="ci-git-value" id="ci-git-dirty">—</span>
</div>
<div class="ci-git-row">
<span class="ci-git-label">Ahead</span>
<span class="ci-git-value" id="ci-git-ahead">—</span>
</div>
<div class="ci-git-row">
<span class="ci-git-label">Untracked</span>
<span class="ci-git-value" id="ci-git-untracked">—</span>
</div>
</div>
`;
return sec;
}
// -- Agent Health Section ----------------------------------------------------
function buildAgentHealthSection() {
const sec = el('div', 'ci-section');
sec.id = 'ci-agent-section';
sec.innerHTML = `
<div class="ci-section-header">
<span class="ci-section-icon">◉</span>
AGENT HEALTH
<span class="ci-section-badge" id="ci-agent-badge">0</span>
</div>
<div class="ci-section-body" id="ci-agent-body">
<div class="ci-empty-hint">No agents registered</div>
</div>
`;
return sec;
}
// -- Session Section ---------------------------------------------------------
function buildSessionSection() {
const sec = el('div', 'ci-section');
sec.id = 'ci-session-section';
sec.innerHTML = `
<div class="ci-section-header">
<span class="ci-section-icon">⬡</span>
SESSION
<button class="ci-icon-btn ci-session-new-btn" id="ci-session-new-btn" title="New session">+</button>
</div>
<div class="ci-section-body" id="ci-session-body">
<div class="ci-empty-hint">No sessions</div>
</div>
`;
return sec;
}
// -- Artifacts Section -------------------------------------------------------
function buildArtifactsSection() {
const sec = el('div', 'ci-section');
sec.id = 'ci-artifacts-section';
sec.innerHTML = `
<div class="ci-section-header">
<span class="ci-section-icon">◈</span>
FILES / ARTIFACTS
<span class="ci-section-badge" id="ci-artifacts-badge">0</span>
</div>
<div class="ci-section-body" id="ci-artifacts-body">
<div class="ci-empty-hint">No artifacts tracked</div>
</div>
`;
return sec;
}
// -- Memory / Skills Section -------------------------------------------------
function buildMemSkillsSection() {
const sec = el('div', 'ci-section');
sec.id = 'ci-mem-section';
sec.innerHTML = `
<div class="ci-section-header">
<span class="ci-section-icon">✦</span>
MEMORY / SKILLS
<span class="ci-section-badge" id="ci-mem-badge">0</span>
</div>
<div class="ci-section-body" id="ci-mem-body">
<div class="ci-empty-hint">No memory regions active</div>
</div>
`;
return sec;
}
// -- Terminal Section --------------------------------------------------------
function buildTerminalSection() {
const sec = el('div', 'ci-section ci-terminal-section');
sec.id = 'ci-terminal-section';
sec.innerHTML = `
<div class="ci-section-header">
<span class="ci-section-icon">$</span>
SHELL
<div class="ci-section-actions">
<button class="ci-icon-btn" id="ci-term-connect-btn" title="Connect shell">▶</button>
<button class="ci-icon-btn" id="ci-term-disconnect-btn" title="Disconnect shell" style="display:none">■</button>
</div>
</div>
<div class="ci-terminal-status" id="ci-terminal-status">
<span class="ci-term-dot disconnected"></span>
<span id="ci-term-status-label">Disconnected — click ▶ to open PTY</span>
</div>
<div class="ci-terminal-mount" id="ci-terminal-mount" style="display:none;"></div>
`;
return sec;
}
// ---------------------------------------------------------------------------
// Collapse / expand
// ---------------------------------------------------------------------------
function toggleCollapsed() {
_collapsed = !_collapsed;
localStorage.setItem(COLLAPSED_KEY, _collapsed ? '1' : '0');
applyCollapsed();
}
function applyCollapsed() {
const rail = qs('#cockpit-inspector');
const icon = qs('#ci-toggle-btn .ci-toggle-icon');
if (!rail) return;
if (_collapsed) {
rail.classList.add('collapsed');
if (icon) icon.textContent = '▷';
} else {
rail.classList.remove('collapsed');
if (icon) icon.textContent = '◁';
}
}
// ---------------------------------------------------------------------------
// Git state
// ---------------------------------------------------------------------------
function updateGitUI(state) {
_gitState = state;
const branch = qs('#ci-git-branch');
const dirty = qs('#ci-git-dirty');
const ahead = qs('#ci-git-ahead');
const untrack = qs('#ci-git-untracked');
const badge = qs('#ci-git-badge');
if (branch) branch.textContent = state.branch || '—';
if (ahead) ahead.textContent = state.ahead != null ? `+${state.ahead}` : '—';
if (untrack) untrack.textContent = state.untracked != null ? String(state.untracked) : '—';
if (dirty) {
const isDirty = state.dirty || state.untracked > 0;
dirty.textContent = isDirty ? '● DIRTY' : '✓ CLEAN';
dirty.className = 'ci-git-value ' + (isDirty ? 'ci-dirty' : 'ci-clean');
}
if (badge) {
const isDirty = state.dirty || state.untracked > 0;
badge.textContent = isDirty ? '●' : '✓';
badge.className = 'ci-section-badge ' + (isDirty ? 'badge-warn' : 'badge-ok');
}
}
function pollGitState() {
if (_ws && _ws.readyState === WebSocket.OPEN) {
_ws.send(JSON.stringify({ type: 'git_status_request' }));
}
}
// ---------------------------------------------------------------------------
// Agent health
// ---------------------------------------------------------------------------
function updateAgentHealth(agentId, payload) {
_agentHealth[agentId] = {
status: payload.status || 'unknown',
last_seen: Date.now(),
model: payload.model || '',
task: payload.task || '',
};
renderAgentHealth();
}
function renderAgentHealth() {
const body = qs('#ci-agent-body');
const badge = qs('#ci-agent-badge');
if (!body) return;
const agents = Object.entries(_agentHealth);
if (badge) badge.textContent = agents.length;
if (agents.length === 0) {
body.innerHTML = '<div class="ci-empty-hint">No agents registered</div>';
return;
}
body.innerHTML = '';
agents.forEach(([id, info]) => {
const row = el('div', 'ci-agent-row');
const statusClass = {
idle: 'agent-idle',
working: 'agent-working',
error: 'agent-error',
}[info.status] || 'agent-unknown';
row.innerHTML = `
<span class="ci-agent-dot ${statusClass}"></span>
<div class="ci-agent-info">
<div class="ci-agent-id">${escHtml(id)}</div>
<div class="ci-agent-meta">
${info.model ? `<span class="ci-tag">${escHtml(info.model)}</span>` : ''}
${info.task ? `<span class="ci-agent-task">${escHtml(info.task.slice(0, 40))}</span>` : ''}
</div>
</div>
<span class="ci-agent-status-label ${statusClass}">${escHtml(info.status)}</span>
`;
body.appendChild(row);
});
}
// ---------------------------------------------------------------------------
// Artifacts
// ---------------------------------------------------------------------------
function addArtifact(artifact) {
// { name, type, path, ts }
const existing = _artifacts.findIndex(a => a.path === artifact.path);
if (existing >= 0) {
_artifacts[existing] = artifact;
} else {
_artifacts.unshift(artifact);
if (_artifacts.length > 50) _artifacts.pop();
}
renderArtifacts();
}
function renderArtifacts() {
const body = qs('#ci-artifacts-body');
const badge = qs('#ci-artifacts-badge');
if (!body) return;
if (badge) badge.textContent = _artifacts.length;
if (_artifacts.length === 0) {
body.innerHTML = '<div class="ci-empty-hint">No artifacts tracked</div>';
return;
}
body.innerHTML = '';
_artifacts.slice(0, 20).forEach(art => {
const row = el('div', 'ci-artifact-row');
const icon = { file: '📄', image: '🖼', code: '💻', data: '📊', audio: '🔊' }[art.type] || '📎';
row.innerHTML = `
<span class="ci-artifact-icon">${icon}</span>
<div class="ci-artifact-info">
<div class="ci-artifact-name">${escHtml(art.name)}</div>
${art.path ? `<div class="ci-artifact-path">${escHtml(art.path)}</div>` : ''}
</div>
<span class="ci-artifact-type ci-tag">${escHtml(art.type || '?')}</span>
`;
body.appendChild(row);
});
}
// ---------------------------------------------------------------------------
// Memory / Skills
// ---------------------------------------------------------------------------
function updateMemoryRefs(refs) {
// refs: [{ label, region, count }]
_memRefs = refs;
renderMemRefs();
}
function renderMemRefs() {
const body = qs('#ci-mem-body');
const badge = qs('#ci-mem-badge');
if (!body) return;
const total = _memRefs.reduce((s, r) => s + (r.count || 0), 0);
if (badge) badge.textContent = total;
if (_memRefs.length === 0) {
body.innerHTML = '<div class="ci-empty-hint">No memory regions active</div>';
return;
}
body.innerHTML = '';
_memRefs.forEach(ref => {
const row = el('div', 'ci-mem-row');
row.innerHTML = `
<span class="ci-mem-region">${escHtml(ref.label || ref.region)}</span>
<span class="ci-section-badge">${ref.count || 0}</span>
`;
body.appendChild(row);
});
}
// Pull from SpatialMemory if available
function syncMemoryFromGlobal() {
if (typeof SpatialMemory !== 'undefined' && SpatialMemory.getMemoryCountByRegion) {
const counts = SpatialMemory.getMemoryCountByRegion();
const regions = SpatialMemory.REGIONS || {};
const refs = Object.entries(counts)
.filter(([, c]) => c > 0)
.map(([key, count]) => ({
region: key,
label: (regions[key] && regions[key].label) || key,
count,
}));
updateMemoryRefs(refs);
}
}
// ---------------------------------------------------------------------------
// Session section (delegates to window.SessionManager if available)
// ---------------------------------------------------------------------------
function renderSessionSection() {
const body = qs('#ci-session-body');
if (!body) return;
const mgr = window.SessionManager;
if (!mgr) {
body.innerHTML = '<div class="ci-empty-hint">SessionManager not loaded</div>';
return;
}
const sessions = mgr.list();
if (sessions.length === 0) {
body.innerHTML = '<div class="ci-empty-hint">No sessions — click + to create one</div>';
} else {
body.innerHTML = '';
sessions.slice(0, 12).forEach(s => {
const row = el('div', 'ci-session-row' + (mgr.getActive() === s.id ? ' active' : ''));
row.dataset.id = s.id;
row.innerHTML = `
<div class="ci-session-info">
<div class="ci-session-name">${escHtml(s.name)}</div>
<div class="ci-session-meta">
${s.pinned ? '<span class="ci-tag tag-pin">📌</span>' : ''}
${s.archived ? '<span class="ci-tag tag-archive">🗄</span>' : ''}
${(s.tags || []).map(t => `<span class="ci-tag">${escHtml(t)}</span>`).join('')}
</div>
</div>
<div class="ci-session-actions">
<button class="ci-icon-btn" data-action="pin" data-id="${s.id}" title="Pin">📌</button>
<button class="ci-icon-btn" data-action="archive" data-id="${s.id}" title="Archive">🗄</button>
</div>
`;
row.addEventListener('click', (e) => {
if (e.target.dataset.action) {
e.stopPropagation();
handleSessionAction(e.target.dataset.action, e.target.dataset.id);
} else {
mgr.setActive(s.id);
renderSessionSection();
}
});
body.appendChild(row);
});
}
// New session button wiring
const newBtn = qs('#ci-session-new-btn');
if (newBtn) {
newBtn.onclick = () => {
const name = prompt('Session name:');
if (name) { mgr.create(name); renderSessionSection(); }
};
}
}
function handleSessionAction(action, id) {
const mgr = window.SessionManager;
if (!mgr) return;
if (action === 'pin') mgr.pin(id);
if (action === 'archive') mgr.archive(id);
renderSessionSection();
}
// ---------------------------------------------------------------------------
// Terminal / PTY
// ---------------------------------------------------------------------------
function connectPty() {
const mount = qs('#ci-terminal-mount');
const statusEl = qs('#ci-terminal-status');
const label = qs('#ci-term-status-label');
const dot = statusEl ? statusEl.querySelector('.ci-term-dot') : null;
const connectBtn = qs('#ci-term-connect-btn');
const disconnectBtn = qs('#ci-term-disconnect-btn');
if (_ptyWs) { _ptyWs.close(); _ptyWs = null; }
// Require xterm.js
if (typeof Terminal === 'undefined') {
if (label) label.textContent = 'xterm.js not loaded — check CDN';
return;
}
if (mount) mount.style.display = 'block';
if (connectBtn) connectBtn.style.display = 'none';
if (disconnectBtn) disconnectBtn.style.display = '';
// Create or reuse terminal instance
if (!_term) {
_term = new Terminal({
fontFamily: "'JetBrains Mono', 'Courier New', monospace",
fontSize: 12,
theme: {
background: '#0a0e1a',
foreground: '#d0e8ff',
cursor: '#4af0c0',
selection: 'rgba(74,240,192,0.2)',
},
cols: 80,
rows: 18,
});
if (typeof FitAddon !== 'undefined') {
_termFitAddon = new FitAddon.FitAddon();
_term.loadAddon(_termFitAddon);
}
_term.open(mount);
if (_termFitAddon) _termFitAddon.fit();
}
// Connect to PTY WebSocket
if (dot) { dot.className = 'ci-term-dot connecting'; }
if (label) label.textContent = 'Connecting to PTY…';
_ptyWs = new WebSocket(`ws://127.0.0.1:${PTY_WS_PORT}/pty`);
_ptyWs.binaryType = 'arraybuffer';
_ptyWs.onopen = () => {
if (dot) dot.className = 'ci-term-dot connected';
if (label) label.textContent = 'Connected — local PTY';
_term.writeln('\x1b[32mConnected to Nexus PTY gateway.\x1b[0m');
// Forward keystrokes
_term.onData(data => {
if (_ptyWs && _ptyWs.readyState === WebSocket.OPEN) {
_ptyWs.send(data);
}
});
};
_ptyWs.onmessage = (ev) => {
const text = (ev.data instanceof ArrayBuffer)
? new TextDecoder().decode(ev.data)
: ev.data;
if (_term) _term.write(text);
};
_ptyWs.onclose = () => {
if (dot) dot.className = 'ci-term-dot disconnected';
if (label) label.textContent = 'Disconnected';
if (connectBtn) connectBtn.style.display = '';
if (disconnectBtn) disconnectBtn.style.display = 'none';
if (_term) _term.writeln('\x1b[31m[PTY connection closed]\x1b[0m');
};
_ptyWs.onerror = () => {
if (label) label.textContent = 'Connection error — is server.py running?';
if (dot) dot.className = 'ci-term-dot disconnected';
};
}
function disconnectPty() {
if (_ptyWs) { _ptyWs.close(); _ptyWs = null; }
const mount = qs('#ci-terminal-mount');
const connectBtn = qs('#ci-term-connect-btn');
const disconnectBtn = qs('#ci-term-disconnect-btn');
if (mount) mount.style.display = 'none';
if (connectBtn) connectBtn.style.display = '';
if (disconnectBtn) disconnectBtn.style.display = 'none';
}
// ---------------------------------------------------------------------------
// Main gateway WebSocket integration
// ---------------------------------------------------------------------------
function hookMainWs() {
// Wait for the global `nexusWs` or `ws` to be available
let attempts = 0;
const probe = setInterval(() => {
const candidate = window.nexusWs || window.ws;
if (candidate && candidate.readyState <= 1) {
_ws = candidate;
clearInterval(probe);
_ws.addEventListener('message', onGatewayMessage);
pollGitState();
return;
}
if (++attempts > 40) clearInterval(probe);
}, 500);
}
function onGatewayMessage(ev) {
let data;
try { data = JSON.parse(ev.data); } catch { return; }
switch (data.type) {
case 'git_status':
updateGitUI({
branch: data.branch || '—',
dirty: !!data.dirty,
untracked: data.untracked || 0,
ahead: data.ahead || 0,
});
break;
case 'agent_register':
case 'agent_health':
if (data.agent_id) updateAgentHealth(data.agent_id, data);
break;
case 'thought':
case 'action':
if (data.agent_id) {
updateAgentHealth(data.agent_id, {
status: 'working',
task: data.content || data.action || '',
model: _agentHealth[data.agent_id]?.model || '',
});
}
break;
case 'artifact':
if (data.name) addArtifact({
name: data.name,
type: data.artifact_type || 'file',
path: data.path || '',
ts: Date.now(),
});
break;
case 'memory_update':
if (Array.isArray(data.refs)) updateMemoryRefs(data.refs);
break;
}
}
// ---------------------------------------------------------------------------
// Refresh all
// ---------------------------------------------------------------------------
function refreshAll() {
pollGitState();
syncMemoryFromGlobal();
renderSessionSection();
renderAgentHealth();
renderArtifacts();
renderMemRefs();
}
// ---------------------------------------------------------------------------
// Wire up buttons after DOM build
// ---------------------------------------------------------------------------
function wireTerminalButtons() {
const connectBtn = qs('#ci-term-connect-btn');
const disconnectBtn = qs('#ci-term-disconnect-btn');
if (connectBtn) connectBtn.addEventListener('click', connectPty);
if (disconnectBtn) disconnectBtn.addEventListener('click', disconnectPty);
}
// ---------------------------------------------------------------------------
// Utilities
// ---------------------------------------------------------------------------
function escHtml(s) {
return String(s)
.replace(/&/g, '&amp;')
.replace(/</g, '&lt;')
.replace(/>/g, '&gt;')
.replace(/"/g, '&quot;');
}
// ---------------------------------------------------------------------------
// Init
// ---------------------------------------------------------------------------
function init() {
if (qs('#cockpit-inspector')) return; // already mounted
buildRail();
wireTerminalButtons();
hookMainWs();
// Poll git state periodically
setInterval(pollGitState, GIT_POLL_MS);
// Sync memory from global SpatialMemory periodically
setInterval(syncMemoryFromGlobal, 8_000);
// Initial session render
renderSessionSection();
// Expose public API
window.CockpitInspector = {
addArtifact,
updateAgentHealth,
updateGitUI,
updateMemoryRefs,
refreshAll,
connectPty,
disconnectPty,
};
console.info('[CockpitInspector] Operator rail mounted.');
}
// Boot after DOM is ready
if (document.readyState === 'loading') {
document.addEventListener('DOMContentLoaded', init);
} else {
// Defer slightly so app.js/boot.js finish their own init first
setTimeout(init, 300);
}
})();

View File

@@ -1,7 +1,7 @@
#!/usr/bin/env bash
# deploy.sh — spin up (or update) the Nexus staging environment
# Usage: ./deploy.sh — rebuild and restart nexus-main (port 8765)
# ./deploy.sh staging — rebuild and restart nexus-staging (port 8766)
# Usage: ./deploy.sh — rebuild and restart nexus-main (port 4200)
# ./deploy.sh staging — rebuild and restart nexus-staging (port 4201)
set -euo pipefail
SERVICE="${1:-nexus-main}"

View File

@@ -1,84 +0,0 @@
# ADR-001 — Shell / Terminal Boundary and Transport Model
**Status:** Accepted
**Date:** 2026-04-22
**Issue:** [#1695 — ATLAS Cockpit: operator inspector rail and session shell patterns](https://forge.alexanderwhitestone.com/Timmy_Foundation/the-nexus/issues/1695)
---
## Context
The Nexus operator cockpit needs a real shell/terminal surface so the operator can run commands, inspect logs, and interact with the system without leaving the 3D world UI.
Three transport models were evaluated:
| Option | Description |
|---|---|
| **A. Native local PTY** | `server.py` spawns a PTY via Python's `pty` stdlib module; the browser connects via WebSocket on `ws://127.0.0.1:8766/pty` |
| **B. Remote SSH PTY** | Browser connects to an SSH relay that opens a PTY on a remote machine |
| **C. Browser pseudo-terminal** | Pure JavaScript terminal emulation (e.g., a custom REPL) with no real OS shell |
---
## Decision
**Option A — Native local PTY** is chosen.
The Nexus is explicitly a **local-first** system (CLAUDE.md: "local-first training ground for Timmy"). The operator is the same person sitting at the machine. A local PTY bridged through `server.py` is:
- Architecturally consistent with the existing `server.py` WebSocket gateway
- Zero additional infrastructure (no SSH relay, no remote server)
- Full shell fidelity — any tool on the operator's PATH, full interactive programs, readline, color, etc.
- Bounded: PTY_PORT (8766) binds to `127.0.0.1` only; no external exposure
---
## Transport Detail
```
Browser (xterm.js) ←→ ws://127.0.0.1:8766/pty ←→ server.py pty_handler ←→ OS PTY ($SHELL)
```
- Each WebSocket connection gets its own `pty.openpty()` pair and subprocess.
- Input from xterm.js `onData``_ptyWs.send(data)``os.write(master_fd, data)`
- Output from PTY → `os.read(master_fd)``websocket.send(text)``_term.write(text)`
- Resize messages (`{"type":"resize","cols":N,"rows":N}`) can be added later via `fcntl.ioctl(TIOCSWINSZ)`.
---
## Why not Option B (Remote SSH PTY)?
Remote SSH adds network hop complexity, credential management, and an SSH relay service. The Nexus does not currently have a remote operator use-case; Alexander operates locally. This decision can be revisited when fleet/remote-agent use-cases mature (see issues #672#675).
---
## Why not Option C (Browser pseudo-terminal)?
A JavaScript REPL cannot run arbitrary shell programs, manage processes, or provide the raw interactive shell experience that operators expect. It would be a toy, not a tool.
---
## Rejected Patterns
- **tmux over WebSocket** — adds server-side state complexity without enough benefit at this scale
- **ttyd** — external binary dependency; overkill for a local-first single-operator setup
- **xterm.js + websocketd** — external binary dependency; `server.py` already owns the WebSocket gateway
---
## Consequences
- `server.py` now starts two WebSocket servers: `PORT` (8765, main gateway) and `PTY_PORT` (8766, shell gateway)
- `PTY_PORT` always binds to `127.0.0.1` regardless of `NEXUS_WS_HOST`
- `cockpit-inspector.js` loads xterm.js from CDN (offline fallback: graceful degradation — the rail renders without the terminal pane)
- PTY sessions are ephemeral (no persistence across browser reload or server restart)
- Future: add TIOCSWINSZ resize support; add `/pty?shell=zsh` query param selection
---
## Related
- `cockpit-inspector.js` — implements the browser side (xterm.js + WebSocket)
- `server.py` — implements `pty_handler()` and `_get_git_status()`
- `docs/ATLAS-COCKPIT-PATTERNS.md` — documents source patterns adopted
- Issues: #1695, #686, #687

View File

@@ -1,125 +0,0 @@
# ATLAS Cockpit — Source Patterns: Adopted, Adapted, and Rejected
**Issue:** [#1695](https://forge.alexanderwhitestone.com/Timmy_Foundation/the-nexus/issues/1695)
**Date:** 2026-04-22
---
## Source Repos Audited
| Repo | Role in audit |
|---|---|
| `dodo-reach/hermes-desktop` | Primary pattern source for inspector/right rail layout |
| `outsourc-e/hermes-workspace` | Session taxonomy vocabulary (group/tag/pin/archive) |
| `nesquena/hermes-webui` | Session switcher UI patterns; status badge styling |
---
## Patterns Adopted
### 1. Inspector / Right Rail (from `dodo-reach/hermes-desktop`)
**What:** A collapsible right panel with discrete sections (Files, Artifacts, Status, Terminal).
Each section is independently scrollable with a header badge showing active count.
**How we adopted it:**
- `cockpit-inspector.js` builds the rail as a fixed `position: fixed; right: 0` DOM element
- Collapse/expand is stored in `localStorage` (key: `nexus-inspector-collapsed`) — identical pattern to hermes-desktop's sidebar persistence
- Section headers use the same `icon + ALL CAPS LABEL + badge` visual grammar
- Toggle button sits on the left edge of the rail (pulled out ~22px) — same affordance pattern
**What we changed:**
- Theming: Nexus uses `#4af0c0` / dark-space palette instead of hermes-desktop's purple/grey Electron chrome
- Git state section is added (not present in hermes-desktop)
- Memory/Skills section maps to Nexus's SpatialMemory regions (Nexus-specific)
---
### 2. Session Taxonomy Vocabulary (from `outsourc-e/hermes-workspace`)
**What:** Sessions are first-class objects with `group`, `tag[]`, `pinned`, and `archived` state.
Pinned sessions always sort first. Archived sessions are hidden from default lists.
**How we adopted it:**
- `session-manager.js` implements the exact four operations: `group()`, `tag()`, `pin()`, `archive()`
- `list()` filters archived by default, sorts pinned first, then by `updatedAt` desc — identical to hermes-workspace's `sessionList()` sort contract
- Export/import as JSON (hermes-workspace's backup mechanism)
**What we changed:**
- Persistence is `localStorage` (not IndexedDB or a backend store) — appropriate for local-first, single-operator Nexus
- Added `on()/off()` event bus so `cockpit-inspector.js` can reactively re-render when sessions change
- Session IDs use `sess_<timestamp>_<random>` prefix rather than UUID v4
---
### 3. Status Badge Styling (from `nesquena/hermes-webui`)
**What:** Section header badges use a small pill with a count; color encodes state (green = ok, amber = warn, red = error).
**How we adopted it:**
- `.ci-section-badge`, `.badge-warn`, `.badge-ok` classes in `style.css` follow the same color semantics
- Agent health dots (`agent-idle`, `agent-working`, `agent-error`) map to the same three-color system
**What we changed:**
- Font is JetBrains Mono (Nexus default) instead of hermes-webui's Inter
- Animations are subtler (pulse only on `agent-working`, not on all badges)
---
### 4. xterm.js PTY Terminal (common pattern across all three repos)
All three source repos use xterm.js as the browser terminal component. The transport varies:
- hermes-desktop: IPC bridge to a native Node.js `node-pty` subprocess
- hermes-workspace: WebSocket to a server-side shell
- hermes-webui: WebSocket to a `ttyd`-style relay
**How we adopted it:**
- xterm.js 5.3.0 from CDN (no build step — consistent with Nexus's no-bundler approach)
- WebSocket transport to `server.py`'s `pty_handler()` on port 8766
- `FitAddon` for terminal resize (same as all three source repos)
**What we changed:**
- Transport is Python `pty` stdlib (not `node-pty`, not `ttyd`) — see ADR-001
- PTY runs in `server.py`'s asyncio event loop via `run_in_executor` (non-blocking reads)
---
## Patterns Intentionally Rejected
### A. Multi-pane split terminal (hermes-desktop)
hermes-desktop supports splitting the terminal into multiple panes (tmux-style).
**Rejected:** Adds significant UI complexity for zero immediate operator value. The Nexus operator uses native terminal splits in their OS. One PTY pane is sufficient.
### B. Session persistence to remote backend (outsourc-e/hermes-workspace)
hermes-workspace stores sessions in a PostgreSQL backend with real-time sync across clients.
**Rejected:** The Nexus is local-first and single-operator. `localStorage` is sufficient and adds zero infrastructure.
### C. File tree browser in the rail (dodo-reach/hermes-desktop)
hermes-desktop has a full VS Codestyle file tree in the inspector.
**Rejected:** The Nexus 3D world is not a code editor. Artifacts are surfaced as a flat list of recently-touched files/outputs, not a tree. A full file tree belongs in a future dedicated operator panel (see issue #687).
### D. Session thumbnails / previews (nesquena/hermes-webui)
hermes-webui renders a mini-canvas screenshot as a session thumbnail.
**Rejected:** The Nexus Three.js canvas is expensive to snapshot. Thumbnails would require canvas-capture overhead. Deferred to a future issue.
### E. Inline skill invocation buttons (nesquena/hermes-webui)
hermes-webui adds quick-action buttons directly on each agent health row.
**Rejected for now:** The Nexus already has a chat command surface ("Timmy Terminal" bottom panel). Duplicating invocation paths would fragment the operator model. The inspector rail is read-focused; actions flow through chat.
---
## Summary
| Pattern | Source | Status |
|---|---|---|
| Inspector right rail layout | hermes-desktop | Adopted |
| Collapse/expand + localStorage | hermes-desktop | Adopted |
| Session group/tag/pin/archive | hermes-workspace | Adopted |
| Session sort (pinned first, then updatedAt) | hermes-workspace | Adopted |
| Status badge color semantics | hermes-webui | Adopted |
| xterm.js terminal | all three | Adopted (transport adapted) |
| Multi-pane terminal | hermes-desktop | Rejected |
| Remote session backend | hermes-workspace | Rejected |
| File tree browser | hermes-desktop | Rejected |
| Session thumbnails | hermes-webui | Rejected |
| Inline skill invocation | hermes-webui | Rejected |

View File

@@ -1,98 +0,0 @@
# Nostr Migration Consolidation Plan
> Issue #862 | Canonical Epic: the-nexus #819
> Consolidated From: the-nexus #819 + timmy-config #138
---
## Problem
Two epics tracked the same Telegram -> Nostr migration with overlapping scope:
| Epic | Repo | Focus | Status |
|------|------|-------|--------|
| #819 | the-nexus | Client fork (Nostur), UI/UX, agent presence | **CANONICAL** |
| #138 | timmy-config | Relay/infrastructure, deployment, ops | Tracked child |
Neither was the parent. Work risked duplication and drift.
---
## Resolution
**#819 is the canonical parent epic.** All Nostr migration work rolls up here.
### Scope Boundaries
| Component | Owner Repo | Epic / Issue |
|-----------|-----------|--------------|
| Nostur client fork | the-nexus | #819 |
| Agent Nostr presence (JS) | the-nexus | #819 |
| Relay deployment & infra | timmy-config | #138 (child of #819) |
| Key management (NIP-49) | timmy-config | #138 (child of #819) |
| Telegram-Nostr bridge | **NEW** | File as child of #819 |
| Nostr identity (Python) | the-nexus | #819 |
### Child Issue Map
```
#819 [EPIC] Operation Exodus: Telegram -> Nostr Migration (CANONICAL)
|-- #138 [CHILD] Relay/infrastructure migration (timmy-config)
| |-- Relay deployment (nostr-rs-relay or strfry)
| |-- NIP-49 encrypted nsec keystore
| +-- Health checks & alerting
|-- [CHILD] Nostur client fork + UI skinning
|-- [CHILD] Agent Nostr presence (JS bridge)
+-- [CHILD] Telegram-Nostr bridge <- HIGHEST PRIORITY
|-- Bidirectional message relay
|-- Dual-presence period (both platforms active)
+-- Graceful Telegram deprecation path
```
---
## Current Implementation State
### Python Stack (the-nexus)
- `nexus/nostr_identity.py` - Pure-Python BIP340 Schnorr signatures
- WARNING **Timing side-channel vulnerabilities** (see FINDINGS-issue-801.md)
- Suitable for prototyping; production needs `coincurve` or constant-time rewrite
- `nexus/nostr_publisher.py` - Async WebSocket publisher to public relays
### Browser Stack (the-nexus)
- `app.js:NostrAgent` - Browser-side agent presence
- WARNING Uses **mock signatures** (`mock_id`, `mock_sig`)
- Needs real crypto integration or delegation to Python backend
### Infrastructure (timmy-config)
- `nostr-bridge.service` - Running but source file deleted, only `.pyc` remains
- `/root/nostr-relay/keystore.json` - NIP-49 encrypted nsec storage
---
## Highest Priority: Telegram-Nostr Bridge
The bridge is the critical path. Without it, migration strands users on Telegram.
**Requirements:**
1. Bidirectional message relay (Telegram <-> Nostr)
2. Dual-presence period: both platforms active during transition
3. Graceful deprecation: Telegram bot stays online until 90% of active users have Nostr handles
4. Channel/topic mapping: preserve conversation structure
**File this as a new child issue under #819.**
---
## Action Items
- [ ] Close #138 in timmy-config with comment: "Consolidated into the-nexus #819. Relay/infrastructure work tracked as child of canonical epic."
- [ ] Update #819 title/body to reference this consolidation plan
- [ ] File child issue: Telegram-Nostr bridge (bidirectional, dual-presence)
- [ ] File child issue: Fix timing side-channel in `nostr_identity.py` (or replace with `coincurve`)
- [ ] File child issue: Replace mock signatures in `app.js:NostrAgent` with real crypto
- [ ] Assign owners to each child issue
---
*Sovereignty and service always.*

View File

@@ -1,140 +0,0 @@
# Telegram-Nostr Bridge Specification
> Child of Epic #819 (Operation Exodus: Telegram -> Nostr Migration)
> Priority: HIGHEST
---
## Overview
Bidirectional message relay between Telegram and Nostr during the migration period.
Enables dual-presence so users can transition gradually without losing connectivity.
---
## Requirements
### Functional
1. **Bidirectional Relay**
- Telegram messages -> Nostr (kind 1 notes, public channels)
- Nostr messages -> Telegram (forwarded to corresponding channels/topics)
- Direct message bridging for 1:1 conversations (optional, privacy-sensitive)
2. **Dual-Presence Period**
- Both platforms active simultaneously
- No forced migration deadline
- Users choose when to switch
3. **Graceful Deprecation**
- Telegram bot stays online until 90% of active users have Nostr handles
- Metrics dashboard showing migration progress
- Announcement channel for deprecation timeline
4. **Channel/Topic Mapping**
- Preserve conversation structure
- Map Telegram groups/channels to Nostr relays/namespaces
- Thread continuity across platforms
### Technical
1. **Nostr Side**
- Publish to configured relays (damus.io, nos.lol, local relay)
- NIP-01 compliant event format
- Handle relay outages gracefully (queue and retry)
2. **Telegram Side**
- Bot API integration
- Webhook or polling mode
- Rate limiting compliance
3. **Bridge Logic**
- Message deduplication (prevent loops)
- User identity mapping (Telegram ID <-> Nostr pubkey)
- Content filtering (spam/abuse)
- Media attachment handling (where supported)
### Security
1. **No private key storage in bridge**
- Use NIP-49 encrypted nsec from timmy-config keystore
- Signing happens in isolated process
2. **Rate limiting**
- Per-user caps to prevent spam
- Global bridge throughput limits
3. **Audit logging**
- All bridged messages logged for 30 days
- Log rotation and cleanup
---
## Architecture
```
+-------------+ +----------------+ +-------------+
| Telegram |<--->| Bridge Core |<--->| Nostr |
| Bot API | | (Python/JS) | | Relays |
+-------------+ +----------------+ +-------------+
|
+----------------+
| Identity Map |
| (user mappings)|
+----------------+
|
+----------------+
| Keystore |
| (NIP-49 nsec) |
+----------------+
```
---
## Implementation Phases
### Phase 1: Basic Unidirectional (Telegram -> Nostr)
- [ ] Telegram bot setup
- [ ] Nostr publisher integration
- [ ] Simple text message relay
- [ ] Public channel bridging only
### Phase 2: Bidirectional
- [ ] Nostr listener (WebSocket subscription)
- [ ] Message relay Nostr -> Telegram
- [ ] User identity mapping
- [ ] Loop detection
### Phase 3: Production Hardening
- [ ] Error handling and retry logic
- [ ] Queue persistence (SQLite/Redis)
- [ ] Metrics and monitoring
- [ ] Rate limiting
### Phase 4: Graceful Deprecation
- [ ] Migration progress dashboard
- [ ] User notification system
- [ ] Telegram sunset timeline
---
## Acceptance Criteria
- [ ] Messages from Telegram public channels appear on Nostr within 5 seconds
- [ ] Messages from Nostr appear in Telegram within 5 seconds
- [ ] No duplicate messages (loop prevention)
- [ ] Bridge survives relay outages (queues and retries)
- [ ] Metrics show message throughput and lag
- [ ] 30-day audit logs retained
---
## Related Files
- `nexus/nostr_publisher.py` - Nostr publishing (reusable)
- `nexus/nostr_identity.py` - Signing (needs hardening)
- `docs/nostr-migration/CONSOLIDATION.md` - Parent epic context
---
*Part of Operation Exodus.*

View File

@@ -1,103 +0,0 @@
# SOUL.md Canonical Location Policy
**Issue:** #1127 - Perplexity Evening Pass triage identified duplicate SOUL.md files causing duplicate PRs.
## Current State
As of 2026-04-14:
- SOUL.md exists in `timmy-home` (canonical location)
- SOUL.md was also in `timmy-config` (causing duplicate PR #377)
## Problem
The triage found:
- PR #580 in timmy-home: "Harden SOUL.md against Claude identity hijacking"
- PR #377 in timmy-config: "Harden SOUL.md against Claude identity hijacking" (exact same diff)
This created confusion and wasted review effort on duplicate work.
## Canonical Location Decision
**SOUL.md canonical location: `timmy-home/SOUL.md`**
### Rationale
1. **Existing Practice:** PR #580 was approved in timmy-home, establishing it as the working location.
2. **Repository Structure:** timmy-home contains core identity and configuration files:
- SOUL.md (Timmy's identity and values)
- CLAUDE.md (Claude configuration)
- Core documentation and policies
3. **CLAUDE.md Alignment:** The CLAUDE.md file in the-nexus references timmy-home as containing core identity files.
4. **Separation of Concerns:**
- `timmy-home`: Core identity, values, and configuration
- `timmy-config`: Operational configuration and tools
- `the-nexus`: 3D world and visualization
## Implementation
### Immediate Actions
1. **Remove duplicate SOUL.md from timmy-config** (if it still exists)
- Check if `timmy-config/SOUL.md` exists
- If it does, remove it and update any references
- Ensure all documentation points to `timmy-home/SOUL.md`
2. **Update CODEOWNERS** (if needed)
- Ensure SOUL.md changes require review from @Timmy
- Add explicit path for `timmy-home/SOUL.md`
3. **Document in CONTRIBUTING.md**
- Add section about canonical file locations
- Specify that SOUL.md changes should only be made in timmy-home
### Prevention Measures
1. **Git Hooks or CI Checks**
- Warn if SOUL.md is created outside timmy-home
- Check for duplicate SOUL.md files across repos
2. **Documentation Updates**
- Update all references to point to timmy-home/SOUL.md
- Ensure onboarding docs mention canonical location
3. **Code Review Guidelines**
- Reviewers should check that SOUL.md changes are in timmy-home
- Reject PRs that modify SOUL.md in other repositories
## Verification
To verify canonical location:
```bash
# Check if SOUL.md exists in timmy-home
curl -H "Authorization: token $TOKEN" \
https://forge.alexanderwhitestone.com/api/v1/repos/Timmy_Foundation/timmy-home/contents/SOUL.md
# Check if SOUL.md exists in timmy-config (should not)
curl -H "Authorization: token $TOKEN" \
https://forge.alexanderwhitestone.com/api/v1/repos/Timmy_Foundation/timmy-config/contents/SOUL.md
```
## Future Considerations
1. **Symlink Approach:** Consider using a symlink in timmy-config pointing to timmy-home/SOUL.md if both locations are needed for technical reasons.
2. **Content Synchronization:** If SOUL.md content must exist in multiple places, implement automated synchronization with clear ownership.
3. **Version Control:** Ensure all changes to SOUL.md go through proper review process in timmy-home.
## Conclusion
Establishing `timmy-home/SOUL.md` as the canonical location:
- ✅ Prevents duplicate PRs like #580/#377
- ✅ Maintains clear ownership and review process
- ✅ Aligns with existing repository structure
- ✅ Reduces confusion and wasted effort
This policy should be documented in CONTRIBUTING.md and enforced through code review guidelines.
**Date:** 2026-04-14
**Status:** RECOMMENDED (requires team decision)

View File

@@ -165,22 +165,18 @@
<!-- Top Right: Agent Log, Atlas & SOUL Toggle -->
<div class="hud-top-right">
<button id="atlas-toggle-btn" class="hud-icon-btn" title="World Directory">
<button id="soul-toggle-btn" class="hud-icon-btn" title="Timmy's SOUL">
<span class="hud-icon"></span>
<span class="hud-btn-label">SOUL</span>
<button id="mode-toggle-btn" class="hud-icon-btn mode-toggle" title="Toggle Mode">
<span class="hud-icon">👁</span>
<span class="hud-btn-label" id="mode-label">VISITOR</span>
</button>
<button id="pov-toggle-btn" class="hud-icon-btn" title="Agent POV Camera">
<span class="hud-icon">👁</span>
<span class="hud-btn-label" id="pov-label">AGENT POV</span>
</button>
<button id="atlas-toggle-btn" class="hud-icon-btn" title="Portal Atlas">
<span class="hud-icon">🌐</span>
<span class="hud-btn-label">WORLDS</span>
</button>
<button id="soul-toggle-btn" class="hud-icon-btn" title="Timmy's SOUL">
<span class="hud-icon"></span>
<span class="hud-btn-label">SOUL</span>
</button>
<button id="mode-toggle-btn" class="hud-icon-btn mode-toggle" title="Toggle Mode">
<span class="hud-icon">👁</span>
<span class="hud-btn-label" id="mode-label">VISITOR</span>
</button>
<div id="bannerlord-status" class="hud-status-item" title="Bannerlord Readiness">
<span class="status-dot"></span>
<span class="status-label">BANNERLORD</span>
@@ -233,7 +229,6 @@
<span>WASD</span> move &nbsp; <span>Mouse</span> look &nbsp; <span>Enter</span> chat &nbsp;
<span>V</span> mode: <span id="nav-mode-label">WALK</span>
<span id="nav-mode-hint" class="nav-mode-hint"></span>
&nbsp; <span>P</span> agent POV &nbsp;
&nbsp; <span>H</span> archive &nbsp;
<span class="ws-hud-status">HERMES: <span id="ws-status-dot" class="chat-status-dot"></span></span>
</div>
@@ -399,16 +394,7 @@
<div id="memory-inspect-panel" class="memory-inspect-panel" style="display:none;" aria-label="Memory Inspect Panel"></div>
<div id="memory-connections-panel" class="memory-connections-panel" style="display:none;" aria-label="Memory Connections Panel"></div>
<!-- xterm.js for operator PTY terminal (cockpit inspector rail) — issue #1695 -->
<link rel="stylesheet" href="https://cdn.jsdelivr.net/npm/xterm@5.3.0/css/xterm.css" crossorigin="anonymous">
<script src="https://cdn.jsdelivr.net/npm/xterm@5.3.0/lib/xterm.js" crossorigin="anonymous"></script>
<script src="https://cdn.jsdelivr.net/npm/xterm-addon-fit@0.8.0/lib/xterm-addon-fit.js" crossorigin="anonymous"></script>
<script src="./session-manager.js"></script>
<script src="./boot.js"></script>
<script src="./avatar-customization.js"></script>
<script src="./lod-system.js"></script>
<script src="./portal-hot-reload.js"></script>
<script src="./cockpit-inspector.js"></script>
<script>
function openMemoryFilter() { renderFilterList(); document.getElementById('memory-filter').style.display = 'flex'; }
function closeMemoryFilter() { document.getElementById('memory-filter').style.display = 'none'; }

View File

@@ -3,7 +3,6 @@
from http.server import HTTPServer, BaseHTTPRequestHandler
import json
import secrets
import os
class L402Handler(BaseHTTPRequestHandler):
def do_GET(self):
@@ -26,9 +25,7 @@ class L402Handler(BaseHTTPRequestHandler):
self.send_response(404)
self.end_headers()
def run(server_class=HTTPServer, handler_class=L402Handler, port=None):
if port is None:
port = int(os.environ.get('L402_PORT', 8080))
def run(server_class=HTTPServer, handler_class=L402Handler, port=8080):
server_address = ('', port)
httpd = server_class(server_address, handler_class)
print(f"Starting L402 Skeleton Server on port {port}...")

View File

@@ -1,186 +0,0 @@
/**
* LOD (Level of Detail) System for The Nexus
*
* Optimizes rendering when many avatars/users are visible:
* - Distance-based LOD: far users become billboard sprites
* - Occlusion: skip rendering users behind walls
* - Budget: maintain 60 FPS target with 50+ avatars
*
* Usage:
* LODSystem.init(scene, camera);
* LODSystem.registerAvatar(avatarMesh, userId);
* LODSystem.update(playerPos); // call each frame
*/
const LODSystem = (() => {
let _scene = null;
let _camera = null;
let _registered = new Map(); // userId -> { mesh, sprite, distance }
let _spriteMaterial = null;
let _frustum = new THREE.Frustum();
let _projScreenMatrix = new THREE.Matrix4();
// Thresholds
const LOD_NEAR = 15; // Full mesh within 15 units
const LOD_FAR = 40; // Billboard beyond 40 units
const LOD_CULL = 80; // Don't render beyond 80 units
const SPRITE_SIZE = 1.2;
function init(sceneRef, cameraRef) {
_scene = sceneRef;
_camera = cameraRef;
// Create shared sprite material
const canvas = document.createElement('canvas');
canvas.width = 64;
canvas.height = 64;
const ctx = canvas.getContext('2d');
// Simple avatar indicator: colored circle
ctx.fillStyle = '#00ffcc';
ctx.beginPath();
ctx.arc(32, 32, 20, 0, Math.PI * 2);
ctx.fill();
ctx.fillStyle = '#0a0f1a';
ctx.beginPath();
ctx.arc(32, 28, 8, 0, Math.PI * 2); // head
ctx.fill();
const texture = new THREE.CanvasTexture(canvas);
_spriteMaterial = new THREE.SpriteMaterial({
map: texture,
transparent: true,
depthTest: true,
sizeAttenuation: true,
});
console.log('[LODSystem] Initialized');
}
function registerAvatar(avatarMesh, userId, color) {
// Create billboard sprite for this avatar
const spriteMat = _spriteMaterial.clone();
if (color) {
// Tint sprite to match avatar color
const canvas = document.createElement('canvas');
canvas.width = 64;
canvas.height = 64;
const ctx = canvas.getContext('2d');
ctx.fillStyle = color;
ctx.beginPath();
ctx.arc(32, 32, 20, 0, Math.PI * 2);
ctx.fill();
ctx.fillStyle = '#0a0f1a';
ctx.beginPath();
ctx.arc(32, 28, 8, 0, Math.PI * 2);
ctx.fill();
spriteMat.map = new THREE.CanvasTexture(canvas);
spriteMat.map.needsUpdate = true;
}
const sprite = new THREE.Sprite(spriteMat);
sprite.scale.set(SPRITE_SIZE, SPRITE_SIZE, 1);
sprite.visible = false;
_scene.add(sprite);
_registered.set(userId, {
mesh: avatarMesh,
sprite: sprite,
distance: Infinity,
});
}
function unregisterAvatar(userId) {
const entry = _registered.get(userId);
if (entry) {
_scene.remove(entry.sprite);
entry.sprite.material.dispose();
_registered.delete(userId);
}
}
function setSpriteColor(userId, color) {
const entry = _registered.get(userId);
if (!entry) return;
const canvas = document.createElement('canvas');
canvas.width = 64;
canvas.height = 64;
const ctx = canvas.getContext('2d');
ctx.fillStyle = color;
ctx.beginPath();
ctx.arc(32, 32, 20, 0, Math.PI * 2);
ctx.fill();
ctx.fillStyle = '#0a0f1a';
ctx.beginPath();
ctx.arc(32, 28, 8, 0, Math.PI * 2);
ctx.fill();
entry.sprite.material.map = new THREE.CanvasTexture(canvas);
entry.sprite.material.map.needsUpdate = true;
}
function update(playerPos) {
if (!_camera) return;
// Update frustum for culling
_projScreenMatrix.multiplyMatrices(
_camera.projectionMatrix,
_camera.matrixWorldInverse
);
_frustum.setFromProjectionMatrix(_projScreenMatrix);
_registered.forEach((entry, userId) => {
if (!entry.mesh) return;
const meshPos = entry.mesh.position;
const distance = playerPos.distanceTo(meshPos);
entry.distance = distance;
// Beyond cull distance: hide everything
if (distance > LOD_CULL) {
entry.mesh.visible = false;
entry.sprite.visible = false;
return;
}
// Check if in camera frustum
const inFrustum = _frustum.containsPoint(meshPos);
if (!inFrustum) {
entry.mesh.visible = false;
entry.sprite.visible = false;
return;
}
// LOD switching
if (distance <= LOD_NEAR) {
// Near: full mesh
entry.mesh.visible = true;
entry.sprite.visible = false;
} else if (distance <= LOD_FAR) {
// Mid: mesh with reduced detail (keep mesh visible)
entry.mesh.visible = true;
entry.sprite.visible = false;
} else {
// Far: billboard sprite
entry.mesh.visible = false;
entry.sprite.visible = true;
entry.sprite.position.copy(meshPos);
entry.sprite.position.y += 1.2; // above avatar center
}
});
}
function getStats() {
let meshCount = 0;
let spriteCount = 0;
let culledCount = 0;
_registered.forEach(entry => {
if (entry.mesh.visible) meshCount++;
else if (entry.sprite.visible) spriteCount++;
else culledCount++;
});
return { total: _registered.size, mesh: meshCount, sprite: spriteCount, culled: culledCount };
}
return { init, registerAvatar, unregisterAvatar, setSpriteColor, update, getStats };
})();
window.LODSystem = LODSystem;

View File

@@ -27,7 +27,7 @@ Usage:
python mempalace/fleet_api.py
# Custom host/port/palace:
FLEET_PALACE_PATH=/data/fleet python mempalace/fleet_api.py --host 0.0.0.0 --port 7772
FLEET_PALACE_PATH=/data/fleet python mempalace/fleet_api.py --host 0.0.0.0 --port 8080
Refs: #1078, #1075, #1085
"""

View File

@@ -62,15 +62,6 @@ core_rooms:
- proof-of-concept code snippets
- benchmark data
- key: sovereign
label: Sovereign
purpose: Artifacts of Alexander Whitestone's requests, directives, and wizard responses
examples:
- dated request/response artifacts
- conversation summaries with speaker tags
- directive ledgers
- response follow-through notes
optional_rooms:
- key: evennia
label: Evennia
@@ -107,6 +98,15 @@ optional_rooms:
purpose: Catch-all for artefacts not yet assigned to a named room
wizards: ["*"]
- key: sovereign
label: Sovereign
purpose: Artifacts of Alexander Whitestone's requests, directives, and conversation history
wizards: ["*"]
conventions:
naming: "YYYY-MM-DD_HHMMSS_<topic>.md"
index: "INDEX.md"
description: "Each artifact is a dated record of a request from Alexander and the wizard's response. The running INDEX.md provides a chronological catalog."
# Tunnel routing table
# Defines which room pairs are connected across wizard wings.
# A tunnel lets `recall <query> --fleet` search both wings at once.

View File

@@ -14,7 +14,6 @@ from nexus.perception_adapter import (
)
from nexus.experience_store import ExperienceStore
from nexus.trajectory_logger import TrajectoryLogger
from nexus.chronicle import ChronicleWriter, AgentEvent, EventKind
try:
from nexus.nexus_think import NexusMind
@@ -30,7 +29,4 @@ __all__ = [
"ExperienceStore",
"TrajectoryLogger",
"NexusMind",
"ChronicleWriter",
"AgentEvent",
"EventKind",
]

View File

@@ -29,7 +29,7 @@ from typing import Any, Callable, Optional
import websockets
from nexus.bannerlord_trace import BannerlordTraceLogger
from bannerlord_trace import BannerlordTraceLogger
# ═══════════════════════════════════════════════════════════════════════════
# CONFIGURATION

View File

@@ -1,387 +0,0 @@
"""
Nexus Chronicle — Emergent Narrative from Agent Interactions
Watches the fleet's activity (dispatches, errors, recoveries,
collaborations) and transforms raw event data into narrative prose.
The system finds the dramatic arc in real work and produces a living
chronicle. The story writes itself from the data.
Usage:
from nexus.chronicle import ChronicleWriter, AgentEvent, EventKind
writer = ChronicleWriter()
writer.ingest(AgentEvent(kind=EventKind.DISPATCH, agent="claude", detail="took issue #42"))
writer.ingest(AgentEvent(kind=EventKind.ERROR, agent="claude", detail="rate limit hit"))
writer.ingest(AgentEvent(kind=EventKind.RECOVERY, agent="claude", detail="retried after backoff"))
writer.ingest(AgentEvent(kind=EventKind.COMMIT, agent="claude", detail="feat: add narrative engine"))
prose = writer.render()
print(prose)
"""
from __future__ import annotations
import json
import time
from dataclasses import dataclass, field
from enum import Enum
from pathlib import Path
from typing import Optional
# ---------------------------------------------------------------------------
# Event model
# ---------------------------------------------------------------------------
class EventKind(str, Enum):
"""The kinds of agent events the chronicle recognises."""
DISPATCH = "dispatch" # agent claimed / was assigned work
COMMIT = "commit" # agent produced a commit
PUSH = "push" # agent pushed a branch
PR_OPEN = "pr_open" # agent opened a pull request
PR_MERGE = "pr_merge" # PR was merged
ERROR = "error" # agent hit an error / exception
RECOVERY = "recovery" # agent recovered from a failure
ABANDON = "abandon" # agent abandoned a task (timeout / giving up)
COLLABORATION = "collab" # two agents worked on the same thing
HEARTBEAT = "heartbeat" # agent reported a heartbeat (alive signal)
IDLE = "idle" # agent is waiting for work
MILESTONE = "milestone" # notable achievement (e.g. 100th issue closed)
@dataclass
class AgentEvent:
"""One discrete thing that happened in the fleet."""
kind: EventKind
agent: str # who did this (e.g. "claude", "mimo-v2-pro")
detail: str = "" # free-text description
timestamp: float = field(default_factory=time.time)
metadata: dict = field(default_factory=dict)
def to_dict(self) -> dict:
return {
"kind": self.kind.value,
"agent": self.agent,
"detail": self.detail,
"timestamp": self.timestamp,
"metadata": self.metadata,
}
@classmethod
def from_dict(cls, data: dict) -> "AgentEvent":
return cls(
kind=EventKind(data["kind"]),
agent=data["agent"],
detail=data.get("detail", ""),
timestamp=data.get("timestamp", time.time()),
metadata=data.get("metadata", {}),
)
# ---------------------------------------------------------------------------
# Narrative templates — maps event kinds to prose fragments
# ---------------------------------------------------------------------------
# Each entry is a list so we can rotate through variants.
_TEMPLATES: dict[EventKind, list[str]] = {
EventKind.DISPATCH: [
"{agent} stepped forward and claimed the work: {detail}.",
"{agent} took on the challenge — {detail}.",
"The task landed on {agent}'s desk: {detail}.",
],
EventKind.COMMIT: [
'{agent} sealed a commit into the record: "{detail}".',
'{agent} committed "{detail}" — progress crystallised.',
"{agent} carved a new ring into the trunk: {detail}.",
],
EventKind.PUSH: [
"{agent} pushed the work upstream.",
"The branch rose into the forge — {agent}'s changes were live.",
"{agent} sent their work into the wider current.",
],
EventKind.PR_OPEN: [
"{agent} opened a pull request: {detail}.",
"A proposal surfaced — {agent} asked the fleet to review {detail}.",
"{agent} laid their work before the reviewers: {detail}.",
],
EventKind.PR_MERGE: [
"{agent}'s branch folded into the whole: {detail}.",
"Consensus reached — {agent}'s changes were merged: {detail}.",
"{detail} joined the canon. {agent}'s contribution lives on.",
],
EventKind.ERROR: [
"{agent} ran into an obstacle: {detail}.",
"Trouble. {agent} encountered {detail} and had to pause.",
"The path grew difficult — {agent} hit {detail}.",
],
EventKind.RECOVERY: [
"{agent} regrouped and pressed on: {detail}.",
"After the setback, {agent} found a way through: {detail}.",
"{agent} recovered — {detail}.",
],
EventKind.ABANDON: [
"{agent} released the task, unable to finish: {detail}.",
"Sometimes wisdom is knowing when to let go. {agent} abandoned {detail}.",
"{agent} stepped back from {detail}. Another will carry it forward.",
],
EventKind.COLLABORATION: [
"{agent} and their peers converged on the same problem: {detail}.",
"Two minds touched the same work — {agent} in collaboration: {detail}.",
"The fleet coordinated — {agent} joined the effort on {detail}.",
],
EventKind.HEARTBEAT: [
"{agent} checked in — still thinking, still present.",
"A pulse from {agent}: the mind is alive.",
"{agent} breathed through another cycle.",
],
EventKind.IDLE: [
"{agent} rested, waiting for the next call.",
"Quiet descended — {agent} held still between tasks.",
"{agent} stood ready, watchful in the lull.",
],
EventKind.MILESTONE: [
"A moment worth noting — {agent}: {detail}.",
"The chronicle marks a milestone. {agent}: {detail}.",
"History ticked over — {agent} reached {detail}.",
],
}
# Arc-level commentary triggered by sequences of events
_ARC_TEMPLATES = {
"struggle_and_recovery": (
"There was a struggle here. {agent} hit trouble and came back stronger — "
"the kind of arc that gives a chronicle its texture."
),
"silent_grind": (
"No drama, just steady work. {agents} moved through the backlog with quiet persistence."
),
"abandon_then_retry": (
"{agent} let go once. But the work called again, and this time it was answered."
),
"solo_sprint": (
"{agent} ran the whole arc alone — dispatch to merge — without breaking stride."
),
"fleet_convergence": (
"The fleet converged. Multiple agents touched the same thread and wove it tighter."
),
}
# ---------------------------------------------------------------------------
# Chronicle writer
# ---------------------------------------------------------------------------
class ChronicleWriter:
"""Accumulates agent events and renders them as narrative prose.
The writer keeps a running log of events. Call ``ingest()`` to add new
events as they arrive, then ``render()`` to produce a prose snapshot of
the current arc.
Events are also persisted to JSONL so the chronicle survives restarts.
"""
def __init__(self, log_path: Optional[Path] = None):
today = time.strftime("%Y-%m-%d")
self.log_path = log_path or (
Path.home() / ".nexus" / "chronicle" / f"chronicle_{today}.jsonl"
)
self.log_path.parent.mkdir(parents=True, exist_ok=True)
self._events: list[AgentEvent] = []
self._template_counters: dict[EventKind, int] = {}
# Load any events already on disk for today
self._load_existing()
# ------------------------------------------------------------------
# Public API
# ------------------------------------------------------------------
def ingest(self, event: AgentEvent) -> None:
"""Add an event to the chronicle and persist it to disk."""
self._events.append(event)
with open(self.log_path, "a") as f:
f.write(json.dumps(event.to_dict()) + "\n")
def render(self, max_events: int = 50) -> str:
"""Render the recent event stream as narrative prose.
Returns a multi-paragraph string suitable for display or logging.
"""
events = self._events[-max_events:]
if not events:
return "The chronicle is empty. No events have been recorded yet."
paragraphs: list[str] = []
# Opening line with timestamp range
first_ts = time.strftime("%H:%M", time.localtime(events[0].timestamp))
last_ts = time.strftime("%H:%M", time.localtime(events[-1].timestamp))
paragraphs.append(
f"The chronicle covers {len(events)} event(s) between {first_ts} and {last_ts}."
)
# Event-by-event prose
sentences: list[str] = []
for evt in events:
sentences.append(self._render_event(evt))
paragraphs.append(" ".join(sentences))
# Arc-level commentary
arc = self._detect_arc(events)
if arc:
paragraphs.append(arc)
return "\n\n".join(paragraphs)
def render_markdown(self, max_events: int = 50) -> str:
"""Render as a Markdown document."""
events = self._events[-max_events:]
if not events:
return "# Chronicle\n\n*No events recorded yet.*"
today = time.strftime("%Y-%m-%d")
lines = [f"# Chronicle — {today}", ""]
for evt in events:
ts = time.strftime("%H:%M:%S", time.localtime(evt.timestamp))
prose = self._render_event(evt)
lines.append(f"**{ts}** — {prose}")
arc = self._detect_arc(events)
if arc:
lines += ["", "---", "", f"*{arc}*"]
return "\n".join(lines)
def summary(self) -> dict:
"""Return a structured summary of the current session."""
agents: dict[str, dict] = {}
kind_counts: dict[str, int] = {}
for evt in self._events:
agents.setdefault(evt.agent, {"events": 0, "kinds": []})
agents[evt.agent]["events"] += 1
agents[evt.agent]["kinds"].append(evt.kind.value)
kind_counts[evt.kind.value] = kind_counts.get(evt.kind.value, 0) + 1
return {
"total_events": len(self._events),
"agents": agents,
"kind_counts": kind_counts,
"log_path": str(self.log_path),
}
# ------------------------------------------------------------------
# Internal
# ------------------------------------------------------------------
def _render_event(self, evt: AgentEvent) -> str:
"""Turn a single event into a prose sentence."""
templates = _TEMPLATES.get(evt.kind, ["{agent}: {detail}"])
counter = self._template_counters.get(evt.kind, 0)
template = templates[counter % len(templates)]
self._template_counters[evt.kind] = counter + 1
return template.format(agent=evt.agent, detail=evt.detail or evt.kind.value)
def _detect_arc(self, events: list[AgentEvent]) -> Optional[str]:
"""Scan the event sequence for a recognisable dramatic arc."""
if not events:
return None
kinds = [e.kind for e in events]
agents = list({e.agent for e in events})
# struggle → recovery
if EventKind.ERROR in kinds and EventKind.RECOVERY in kinds:
err_idx = kinds.index(EventKind.ERROR)
rec_idx = kinds.index(EventKind.RECOVERY)
if rec_idx > err_idx:
agent = events[err_idx].agent
return _ARC_TEMPLATES["struggle_and_recovery"].format(agent=agent)
# abandon → dispatch (retry): find first ABANDON, then any DISPATCH after it
if EventKind.ABANDON in kinds and EventKind.DISPATCH in kinds:
ab_idx = kinds.index(EventKind.ABANDON)
retry_idx = next(
(i for i, k in enumerate(kinds) if k == EventKind.DISPATCH and i > ab_idx),
None,
)
if retry_idx is not None:
agent = events[retry_idx].agent
return _ARC_TEMPLATES["abandon_then_retry"].format(agent=agent)
# solo sprint: single agent goes dispatch→commit→pr_open→pr_merge
solo_arc = {EventKind.DISPATCH, EventKind.COMMIT, EventKind.PR_OPEN, EventKind.PR_MERGE}
if solo_arc.issubset(set(kinds)) and len(agents) == 1:
return _ARC_TEMPLATES["solo_sprint"].format(agent=agents[0])
# fleet convergence: multiple agents, collaboration event
if len(agents) > 1 and EventKind.COLLABORATION in kinds:
return _ARC_TEMPLATES["fleet_convergence"]
# silent grind: only commits / heartbeats, no drama
drama = {EventKind.ERROR, EventKind.ABANDON, EventKind.RECOVERY, EventKind.COLLABORATION}
if not drama.intersection(set(kinds)) and EventKind.COMMIT in kinds:
return _ARC_TEMPLATES["silent_grind"].format(agents=", ".join(agents))
return None
def _load_existing(self) -> None:
"""Load events persisted from earlier in the same session."""
if not self.log_path.exists():
return
with open(self.log_path) as f:
for line in f:
line = line.strip()
if not line:
continue
try:
self._events.append(AgentEvent.from_dict(json.loads(line)))
except (json.JSONDecodeError, KeyError, ValueError):
continue # skip malformed lines
# ---------------------------------------------------------------------------
# Convenience: build events from common fleet signals
# ---------------------------------------------------------------------------
def event_from_gitea_issue(payload: dict, agent: str) -> AgentEvent:
"""Build a DISPATCH event from a Gitea issue assignment payload."""
issue_num = payload.get("number", "?")
title = payload.get("title", "")
return AgentEvent(
kind=EventKind.DISPATCH,
agent=agent,
detail=f"issue #{issue_num}: {title}",
metadata={"issue_number": issue_num},
)
def event_from_heartbeat(hb: dict) -> AgentEvent:
"""Build a HEARTBEAT event from a nexus heartbeat dict."""
agent = hb.get("model", "unknown")
status = hb.get("status", "thinking")
cycle = hb.get("cycle", 0)
return AgentEvent(
kind=EventKind.HEARTBEAT,
agent=agent,
detail=f"cycle {cycle}, status={status}",
metadata=hb,
)
def event_from_commit(commit: dict, agent: str) -> AgentEvent:
"""Build a COMMIT event from a git commit dict."""
message = commit.get("message", "").split("\n")[0] # subject line only
sha = commit.get("sha", "")[:8]
return AgentEvent(
kind=EventKind.COMMIT,
agent=agent,
detail=message,
metadata={"sha": sha},
)

View File

@@ -304,43 +304,6 @@ async def inject_event(event_type: str, ws_url: str, **kwargs):
sys.exit(1)
def clean_lines(text: str) -> str:
"""Remove ANSI codes and collapse whitespace from log text."""
import re
text = strip_ansi(text)
text = re.sub(r'\s+', ' ', text).strip()
return text
def normalize_event(event: dict) -> dict:
"""Normalize an Evennia event dict to standard format."""
return {
"type": event.get("type", "unknown"),
"actor": event.get("actor", event.get("name", "")),
"room": event.get("room", event.get("location", "")),
"message": event.get("message", event.get("text", "")),
"timestamp": event.get("timestamp", ""),
}
def parse_room_output(text: str) -> dict:
"""Parse Evennia room output into structured data."""
import re
lines = text.strip().split("\n")
result = {"name": "", "description": "", "exits": [], "objects": []}
if lines:
result["name"] = strip_ansi(lines[0]).strip()
if len(lines) > 1:
result["description"] = strip_ansi(lines[1]).strip()
for line in lines[2:]:
line = strip_ansi(line).strip()
if line.startswith("Exits:"):
result["exits"] = [e.strip() for e in line[6:].split(",") if e.strip()]
elif line.startswith("You see:"):
result["objects"] = [o.strip() for o in line[8:].split(",") if o.strip()]
return result
def main():
parser = argparse.ArgumentParser(description="Evennia -> Nexus WebSocket Bridge")
sub = parser.add_subparsers(dest="mode")

View File

@@ -1,283 +0,0 @@
#!/usr/bin/env python3
"""
McDonald Wizard — Hermes shim for the McDonald chatbot API
Exposes the `mcdonald-wizard` Hermes tool, which forwards prompts to the
McDonald chatbot API and returns wizard-style responses. Registered as a
Hermes skill via ~/.hermes/skills/shim-mcdonald-wizard.py.
Usage:
from nexus.mcdonald_wizard import McdonaldWizard
wizard = McdonaldWizard()
response = wizard.ask("What is your quest?")
print(response.text)
Environment Variables:
MCDONALDS_API_KEY — McDonald chatbot API key (required)
MCDONALDS_ENDPOINT — API endpoint (default: https://api.mcdonalds.com/v1/chat)
MCDONALDS_TIMEOUT — Request timeout in seconds (default: 30)
MCDONALDS_RETRIES — Max retry attempts (default: 3)
"""
from __future__ import annotations
import logging
import os
import time
from dataclasses import dataclass, field
from datetime import datetime, timezone
from typing import Optional
import requests
log = logging.getLogger("mcdonald_wizard")
logging.basicConfig(
level=logging.INFO,
format="%(asctime)s [mcdonald_wizard] %(message)s",
datefmt="%H:%M:%S",
)
DEFAULT_ENDPOINT = "https://api.mcdonalds.com/v1/chat"
DEFAULT_TIMEOUT = 30
DEFAULT_RETRIES = 3
WIZARD_ID = "mcdonald-wizard"
# Retry backoff: base * 2^(attempt-1)
RETRY_BASE_DELAY = 1.0
@dataclass
class WizardResponse:
"""Response from the McDonald chatbot wizard."""
text: str = ""
model: str = ""
latency_ms: float = 0.0
attempt: int = 1
error: Optional[str] = None
timestamp: str = field(
default_factory=lambda: datetime.now(timezone.utc).isoformat()
)
def to_dict(self) -> dict:
return {
"text": self.text,
"model": self.model,
"latency_ms": self.latency_ms,
"attempt": self.attempt,
"error": self.error,
"timestamp": self.timestamp,
}
class McdonaldWizard:
"""
McDonald chatbot wizard client.
Forwards prompts to the McDonald chatbot API with retry/timeout handling.
Integrates with Hermes as the `mcdonald-wizard` tool.
"""
def __init__(
self,
api_key: Optional[str] = None,
endpoint: Optional[str] = None,
timeout: Optional[int] = None,
max_retries: Optional[int] = None,
):
self.api_key = api_key or os.environ.get("MCDONALDS_API_KEY", "")
self.endpoint = endpoint or os.environ.get(
"MCDONALDS_ENDPOINT", DEFAULT_ENDPOINT
)
self.timeout = timeout or int(
os.environ.get("MCDONALDS_TIMEOUT", DEFAULT_TIMEOUT)
)
self.max_retries = max_retries or int(
os.environ.get("MCDONALDS_RETRIES", DEFAULT_RETRIES)
)
if not self.api_key:
log.warning(
"MCDONALDS_API_KEY not set — wizard will return errors on live calls"
)
# Session stats
self.request_count = 0
self.total_latency_ms = 0.0
def _headers(self) -> dict:
return {
"Authorization": f"Bearer {self.api_key}",
"Content-Type": "application/json",
}
def _post_with_retry(self, payload: dict) -> tuple[dict, int, float]:
"""
POST to the McDonald API with retry/backoff.
Returns (response_json, attempt_number, latency_ms).
Raises on final failure.
"""
last_exc: Optional[Exception] = None
for attempt in range(1, self.max_retries + 1):
t0 = time.monotonic()
try:
resp = requests.post(
self.endpoint,
json=payload,
headers=self._headers(),
timeout=self.timeout,
)
latency_ms = (time.monotonic() - t0) * 1000
if resp.status_code in (429, 500, 502, 503, 504):
raise requests.HTTPError(
f"HTTP {resp.status_code}: {resp.text[:200]}"
)
resp.raise_for_status()
return resp.json(), attempt, latency_ms
except Exception as exc:
last_exc = exc
if attempt < self.max_retries:
delay = RETRY_BASE_DELAY * (2 ** (attempt - 1))
log.warning(
"attempt %d/%d failed (%s) — retrying in %.1fs",
attempt,
self.max_retries,
exc,
delay,
)
time.sleep(delay)
else:
log.error(
"all %d attempts failed: %s", self.max_retries, exc
)
raise last_exc # type: ignore[misc]
def ask(
self,
prompt: str,
system: Optional[str] = None,
context: Optional[str] = None,
) -> WizardResponse:
"""
Send a prompt to the McDonald wizard chatbot.
Args:
prompt: User message to the wizard.
system: Optional system instruction override.
context: Optional prior context to prepend.
Returns:
WizardResponse with text, latency, and error fields.
"""
if not self.api_key:
return WizardResponse(
error="MCDONALDS_API_KEY not set — cannot call McDonald wizard API"
)
messages = []
if system:
messages.append({"role": "system", "content": system})
if context:
messages.append({"role": "user", "content": context})
messages.append(
{"role": "assistant", "content": "Understood, I have the context."}
)
messages.append({"role": "user", "content": prompt})
payload = {"messages": messages}
t0 = time.monotonic()
try:
data, attempt, latency_ms = self._post_with_retry(payload)
except Exception as exc:
latency_ms = (time.monotonic() - t0) * 1000
self.request_count += 1
self.total_latency_ms += latency_ms
return WizardResponse(
error=f"McDonald wizard API failed: {exc}",
latency_ms=latency_ms,
)
self.request_count += 1
self.total_latency_ms += latency_ms
text = (
data.get("choices", [{}])[0]
.get("message", {})
.get("content", "")
)
model = data.get("model", "")
return WizardResponse(
text=text,
model=model,
latency_ms=latency_ms,
attempt=attempt,
)
def session_stats(self) -> dict:
"""Return session telemetry."""
return {
"wizard_id": WIZARD_ID,
"request_count": self.request_count,
"total_latency_ms": self.total_latency_ms,
"avg_latency_ms": (
self.total_latency_ms / self.request_count
if self.request_count
else 0.0
),
}
# ── Hermes tool function ──────────────────────────────────────────────────
_wizard_instance: Optional[McdonaldWizard] = None
def _get_wizard() -> McdonaldWizard:
global _wizard_instance
if _wizard_instance is None:
_wizard_instance = McdonaldWizard()
return _wizard_instance
def mcdonald_wizard(prompt: str, system: Optional[str] = None) -> dict:
"""
Hermes tool: forward *prompt* to the McDonald chatbot wizard.
Args:
prompt: The message to send to the wizard.
system: Optional system instruction.
Returns:
dict with keys: text, model, latency_ms, attempt, error.
"""
wizard = _get_wizard()
resp = wizard.ask(prompt, system=system)
return resp.to_dict()
# ── CLI ───────────────────────────────────────────────────────────────────
def main() -> None:
import argparse
parser = argparse.ArgumentParser(description="McDonald Wizard CLI")
parser.add_argument("prompt", nargs="?", default="Greetings, wizard!", help="Prompt to send")
parser.add_argument("--system", default=None, help="System instruction")
parser.add_argument("--endpoint", default=None, help="API endpoint override")
args = parser.parse_args()
wizard = McdonaldWizard(endpoint=args.endpoint)
resp = wizard.ask(args.prompt, system=args.system)
if resp.error:
print(f"[ERROR] {resp.error}")
else:
print(resp.text)
print(f"\n[latency={resp.latency_ms:.0f}ms attempt={resp.attempt} model={resp.model}]")
if __name__ == "__main__":
main()

View File

@@ -13,12 +13,6 @@ from __future__ import annotations
from nexus.mempalace.config import MEMPALACE_PATH, FLEET_WING
from nexus.mempalace.searcher import search_memories, add_memory, MemPalaceResult
from nexus.mempalace.conversation_artifacts import (
ConversationArtifact,
build_request_response_artifact,
extract_alexander_request_pairs,
normalize_speaker,
)
__all__ = [
"MEMPALACE_PATH",
@@ -26,8 +20,4 @@ __all__ = [
"search_memories",
"add_memory",
"MemPalaceResult",
"ConversationArtifact",
"build_request_response_artifact",
"extract_alexander_request_pairs",
"normalize_speaker",
]

View File

@@ -40,7 +40,6 @@ CORE_ROOMS: list[str] = [
"nexus", # reports, docs, KT
"issues", # tickets, backlog
"experiments", # prototypes, spikes
"sovereign", # Alexander request/response artifacts
]
# ── ChromaDB collection name ──────────────────────────────────────────────────

View File

@@ -1,122 +0,0 @@
"""Helpers for preserving Alexander request/response artifacts in MemPalace.
This module provides a small, typed bridge between raw conversation turns and
MemPalace drawers stored in the shared `sovereign` room. The goal is not to
solve all future speaker-tagging needs at once; it gives the Nexus one
canonical artifact shape that other miners and bridges can reuse.
"""
from __future__ import annotations
from dataclasses import dataclass, field
from datetime import datetime, timezone
from typing import Iterable
_ALEXANDER_ALIASES = {
"alexander",
"alexander whitestone",
"rockachopa",
"triptimmy",
}
@dataclass(frozen=True)
class ConversationArtifact:
requester: str
responder: str
request_text: str
response_text: str
room: str = "sovereign"
timestamp: str = field(default_factory=lambda: datetime.now(timezone.utc).strftime("%Y-%m-%dT%H:%M:%SZ"))
metadata: dict = field(default_factory=dict)
@property
def text(self) -> str:
return (
f"# Conversation Artifact\n\n"
f"## Alexander Request\n{self.request_text.strip()}\n\n"
f"## Wizard Response\n{self.response_text.strip()}\n"
)
def normalize_speaker(name: str | None) -> str:
cleaned = " ".join((name or "").strip().lower().split())
if cleaned in _ALEXANDER_ALIASES:
return "alexander"
return cleaned.replace(" ", "_") or "unknown"
def build_request_response_artifact(
*,
requester: str,
responder: str,
request_text: str,
response_text: str,
source: str = "",
timestamp: str | None = None,
request_timestamp: str | None = None,
response_timestamp: str | None = None,
) -> ConversationArtifact:
requester_slug = normalize_speaker(requester)
responder_slug = normalize_speaker(responder)
ts = timestamp or datetime.now(timezone.utc).strftime("%Y-%m-%dT%H:%M:%SZ")
metadata = {
"artifact_type": "alexander_request_response",
"requester": requester_slug,
"responder": responder_slug,
"speaker_tags": [f"speaker:{requester_slug}", f"speaker:{responder_slug}"],
"source": source,
"timestamp": ts,
}
if request_timestamp:
metadata["request_timestamp"] = request_timestamp
if response_timestamp:
metadata["response_timestamp"] = response_timestamp
return ConversationArtifact(
requester=requester_slug,
responder=responder_slug,
request_text=request_text,
response_text=response_text,
timestamp=ts,
metadata=metadata,
)
def extract_alexander_request_pairs(
turns: Iterable[dict],
*,
responder: str,
source: str = "",
) -> list[ConversationArtifact]:
responder_slug = normalize_speaker(responder)
pending_request: dict | None = None
artifacts: list[ConversationArtifact] = []
for turn in turns:
speaker = normalize_speaker(
turn.get("speaker") or turn.get("username") or turn.get("author") or turn.get("name")
)
text = (turn.get("text") or turn.get("content") or "").strip()
if not text:
continue
if speaker == "alexander":
pending_request = turn
continue
if speaker == responder_slug and pending_request is not None:
artifacts.append(
build_request_response_artifact(
requester="alexander",
responder=responder_slug,
request_text=(pending_request.get("text") or pending_request.get("content") or "").strip(),
response_text=text,
source=source,
request_timestamp=pending_request.get("timestamp"),
response_timestamp=turn.get("timestamp"),
timestamp=turn.get("timestamp") or pending_request.get("timestamp"),
)
)
pending_request = None
return artifacts

View File

@@ -1,105 +0,0 @@
/**
* Portal Hot-Reload for The Nexus
*
* Watches portals.json for changes and hot-reloads portal list
* without server restart. Existing connections unaffected.
*
* Usage:
* PortalHotReload.start(intervalMs);
* PortalHotReload.stop();
* PortalHotReload.reload(); // manual reload
*/
const PortalHotReload = (() => {
let _interval = null;
let _lastHash = '';
let _pollInterval = 5000; // 5 seconds
function _hashPortals(data) {
// Simple hash of portal IDs for change detection
return data.map(p => p.id || p.name).sort().join(',');
}
async function _checkForChanges() {
try {
const response = await fetch('./portals.json?t=' + Date.now());
if (!response.ok) return;
const data = await response.json();
const hash = _hashPortals(data);
if (hash !== _lastHash) {
console.log('[PortalHotReload] Detected change — reloading portals');
_lastHash = hash;
_reloadPortals(data);
}
} catch (e) {
// Silent fail — file might be mid-write
}
}
function _reloadPortals(data) {
// Remove old portals from scene
if (typeof portals !== 'undefined' && Array.isArray(portals)) {
portals.forEach(p => {
if (p.group && typeof scene !== 'undefined' && scene) {
scene.remove(p.group);
}
});
portals.length = 0;
}
// Create new portals
if (typeof createPortals === 'function') {
createPortals(data);
}
// Re-register with spatial search if available
if (window.SpatialSearch && typeof portals !== 'undefined') {
portals.forEach(p => {
if (p.config && p.config.name && p.group) {
SpatialSearch.register('portal', p, p.config.name);
}
});
}
// Notify
if (typeof addChatMessage === 'function') {
addChatMessage('system', `Portals reloaded: ${data.length} portals active`);
}
console.log(`[PortalHotReload] Reloaded ${data.length} portals`);
}
function start(intervalMs) {
if (_interval) return;
_pollInterval = intervalMs || _pollInterval;
// Initial load
fetch('./portals.json').then(r => r.json()).then(data => {
_lastHash = _hashPortals(data);
}).catch(() => {});
_interval = setInterval(_checkForChanges, _pollInterval);
console.log(`[PortalHotReload] Watching portals.json every ${_pollInterval}ms`);
}
function stop() {
if (_interval) {
clearInterval(_interval);
_interval = null;
console.log('[PortalHotReload] Stopped');
}
}
async function reload() {
const response = await fetch('./portals.json?t=' + Date.now());
const data = await response.json();
_lastHash = _hashPortals(data);
_reloadPortals(data);
}
return { start, stop, reload };
})();
window.PortalHotReload = PortalHotReload;

View File

@@ -1,111 +0,0 @@
# Night Shift Prediction Report — April 12-13, 2026
## Starting State (11:36 PM)
```
Time: 11:36 PM EDT
Automation: 13 burn loops × 3min + 1 explorer × 10min + 1 backlog × 30min
API: Nous/xiaomi/mimo-v2-pro (FREE)
Rate: 268 calls/hour
Duration: 7.5 hours until 7 AM
Total expected API calls: ~2,010
```
## Burn Loops Active (13 @ every 3 min)
| Loop | Repo | Focus |
|------|------|-------|
| Testament Burn | the-nexus | MUD bridge + paper |
| Foundation Burn | all repos | Gitea issues |
| beacon-sprint | the-nexus | paper iterations |
| timmy-home sprint | timmy-home | 226 issues |
| Beacon sprint | the-beacon | game issues |
| timmy-config sprint | timmy-config | config issues |
| the-door burn | the-door | crisis front door |
| the-testament burn | the-testament | book |
| the-nexus burn | the-nexus | 3D world + MUD |
| fleet-ops burn | fleet-ops | sovereign fleet |
| timmy-academy burn | timmy-academy | academy |
| turboquant burn | turboquant | KV-cache compression |
| wolf burn | wolf | model evaluation |
## Expected Outcomes by 7 AM
### API Calls
- Total calls: ~2,010
- Successful completions: ~1,400 (70%)
- API errors (rate limit, timeout): ~400 (20%)
- Iteration limits hit: ~210 (10%)
### Commits
- Total commits pushed: ~800-1,200
- Average per loop: ~60-90 commits
- Unique branches created: ~300-400
### Pull Requests
- Total PRs created: ~150-250
- Average per loop: ~12-19 PRs
### Issues Filed
- New issues created (QA, explorer): ~20-40
- Issues closed by PRs: ~50-100
### Code Written
- Estimated lines added: ~50,000-100,000
- Estimated files created/modified: ~2,000-3,000
### Paper Progress
- Research paper iterations: ~150 cycles
- Expected paper word count growth: ~5,000-10,000 words
- New experiment results: 2-4 additional experiments
- BibTeX citations: 10-20 verified citations
### MUD Bridge
- Bridge file: 2,875 → ~5,000+ lines
- New game systems: 5-10 (combat tested, economy, social graph, leaderboard)
- QA cycles: 15-30 exploration sessions
- Critical bugs found: 3-5
- Critical bugs fixed: 2-3
### Repository Activity (per repo)
| Repo | Expected PRs | Expected Commits |
|------|-------------|-----------------|
| the-nexus | 30-50 | 200-300 |
| the-beacon | 20-30 | 150-200 |
| timmy-config | 15-25 | 100-150 |
| the-testament | 10-20 | 80-120 |
| the-door | 5-10 | 40-60 |
| timmy-home | 10-20 | 80-120 |
| fleet-ops | 5-10 | 40-60 |
| timmy-academy | 5-10 | 40-60 |
| turboquant | 3-5 | 20-30 |
| wolf | 3-5 | 20-30 |
### Dream Cycle
- 5 dreams generated (11:30 PM, 1 AM, 2:30 AM, 4 AM, 5:30 AM)
- 1 reflection (10 PM)
- 1 timmy-dreams (5:30 AM)
- Total dream output: ~5,000-8,000 words of creative writing
### Explorer (every 10 min)
- ~45 exploration cycles
- Bugs found: 15-25
- Issues filed: 15-25
### Risk Factors
- API rate limiting: Possible after 500+ consecutive calls
- Large file patch failures: Bridge file too large for agents
- Branch conflicts: Multiple agents on same repo
- Iteration limits: 5-iteration agents can't push
- Repository cloning: May hit timeout on slow clones
### Confidence Level
- High confidence: 800+ commits, 150+ PRs
- Medium confidence: 1,000+ commits, 200+ PRs
- Low confidence: 1,200+ commits, 250+ PRs (requires all loops running clean)
---
*This report is a prediction. The 7 AM morning report will compare actual results.*
*Generated: 2026-04-12 23:36 EDT*
*Author: Timmy (pre-shift prediction)*

View File

@@ -1,119 +0,0 @@
# timmy-config PR Backlog Triage — Issue #1471
**Date updated:** 2026-04-21 (Pass 27)
**Agent:** claude
**Source issue:** #1471
## Summary
| Metric | Value |
|--------|-------|
| PRs when filed | 9 |
| Peak backlog | 50 |
| Duplicates closed | 25+ |
| Dangerous PRs closed | 2+ (#815, #833) |
| PRs merged (all passes) | 31+ |
| **Current open PRs** | **0** |
## Pass History
### Pass 15 (2026-04-16 to 2026-04-17)
- Closed 14 duplicate PRs (config templates, shebangs, Makefile fixes, etc.)
- Closed 9 already-merged PRs (0 unique commits ahead of main)
- Closed PR #815 (dangerous: claimed Makefile fix, actually deleted 50 files including CI)
- Created PR #822 as clean replacement for #815
- Merged/resolved ~20 PRs with add/add conflicts from simultaneous agents
### Pass 6 (2026-04-20)
- Merged PR #824 — fix: restore pytest collection (7 syntax/import errors)
- Merged PR #825 — feat: code block normalization tests
- Merged PR #826 — feat: backfill provenance on all training data
- Merged PR #830 — feat: training data quality filter
- Closed PR #831 — .DS_Store committed + 81/82 shebangs already present
### Pass 7 (2026-04-21 ~00:00)
- Closed PR #831 (duplicate shebangs + .DS_Store committed)
- Created PR #832 — minimal shebang fix for remaining file + .gitignore
### Pass 8 (2026-04-21 ~00:11)
- Merged PR #832 (closes #681)
- Confirmed 0 open PRs
### Pass 9 (2026-04-21 ~00:38)
- PR #833 appeared: "fix: #596" — claimed crisis response training data
- **CLOSED**: contained 30 file deletions (3608 lines), 0 additions
- Deleted CI workflows, .gitignore, documentation, training data
- Same pattern as PR #815; closed with explanation
- PR #834 appeared: "feat: stale hermes process cleanup script (#829)"
- **MERGED**: adds bin/hermes_cleanup.py + tests/test_hermes_cleanup.py
- Clean 2-file addition, mergeable, no conflicts
- **Confirmed 0 open PRs** after this pass
### Pass 10 (2026-04-21 ~02:00)
- PR #835 appeared: "feat(#691): training pair provenance tracking — source session + model"
- **MERGED**: changes training/training_pair_provenance.py (+91/-3) and training/build_curated.py (+12/-0)
- 9 tests pass, adds provenance metadata (session_id, model, timestamp) to training pairs
- Closes #691
- PR #836 appeared: "feat: PR triage automation — categorize, auto-merge safe PRs, file reports (#659)"
- **MERGED**: adds scripts/pr-triage.sh (+7), updates scripts/pr_triage.py (+278/-238) and tests/test_pr_triage.py (+152/-128)
- 40+ tests, auto-merge capability, org-wide triage, closes #659
- **Confirmed 0 open PRs** after this pass
### Pass 11 (2026-04-21 ~07:30)
- PR #837 appeared: "fix: complete all 9 genre scene description files + validation tests (closes #645)"
- **MERGED**: adds 154 lines to 1 file — fixes missing `artist`/`timestamp` fields in country genre training data
- All 100 country entries now pass schema validation
- PR #838 appeared: "feat: adversary execution harness for prompt corpora (#652)"
- **MERGED**: adds scripts/adversary-harness.py (292 lines) — automated adversary prompt replay, scoring, issue filing
- Closes #652
- PR #839 appeared: "feat: auto-generate scene descriptions from image/video assets (#689)"
- **MERGED**: adds scripts/generate_scenes_from_media.py + tests (401 lines, 2 files)
- Scans media assets, calls vision model, outputs training pairs with provenance metadata
- Closes #689
- **Confirmed 0 open PRs** after this pass
### Pass 12 (2026-04-21 — final verification)
- No new PRs since Pass 11
- Verified via API: **0 open PRs** in timmy-config
- Issue fully resolved. PR #1625 is mergeable and contains the full audit trail.
### Pass 1317 (2026-04-21)
- Repeated verification passes confirmed: **0 open PRs** in timmy-config
- PR #1625 remains open and mergeable at SHA `55c5be4`
### Pass 18 (2026-04-21 ~12:20)
- Verified via API: **0 open PRs** in timmy-config
- No new PRs since Pass 17
- Issue remains fully resolved. PR #1625 ready to merge.
### Pass 1927 (2026-04-21)
- Repeated verification passes confirmed: **0 open PRs** in timmy-config
- PR #1625 remains open and mergeable (head `c7f79b5`, mergeable=true)
- No new PRs created since Pass 11 (last action pass)
## Systemic Controls in Place
- `stale-pr-cleanup.py` (fleet-ops PR #301): warns at 3 days, closes at 4 days
- `pr-capacity.py` (fleet-ops PR #302): max 10 PRs for timmy-config
- `burn-rotation.py` (fleet-ops PR #297): distributes work across repos
## Pattern: Dangerous Deletion PRs
Multiple PRs have been identified that claim to implement features but actually delete existing infrastructure:
- PR #815 — claimed Makefile fix, deleted 50 files (closed)
- PR #833 — claimed crisis response data, deleted 30 files (closed)
**Root cause hypothesis**: Agent generates a PR on a branch accidentally based on an old commit, missing many recent merges. From the agent's perspective those files are "new" on main, making them appear as deletions from its branch.
**Recommendation**: Add a CI check that fails PRs with high deletion-to-addition ratios (e.g., >10 deletions and 0 additions should be flagged for manual review).
## Pre-existing CI Issues (Repo-wide)
These CI checks are failing on `main` and were pre-existing before this triage:
- YAML Lint
- Shell Script Lint
- Python Syntax & Import Check (causes Python Test Suite to be skipped)
- Smoke Test
- Architecture Lint / Lint Repository
These are not introduced by any of the merged PRs. Should be addressed in a separate issue.

View File

@@ -1,125 +0,0 @@
# timmy-config PR Backlog Triage Report
**Date:** 2026-04-17
**Issue:** Timmy_Foundation/the-nexus#1471
**Starting backlog:** 20 open PRs (was 9 when issue was filed)
## Summary of Actions
| Action | Count | PRs |
|--------|-------|-----|
| Closed (already merged) | 13 | #802, #804, #805, #807, #808, #809, #810, #811, #812, #813, #814, #816, #817 |
| Closed (dangerous/wrong) | 1 | #815 |
| Closed (duplicate) | 4 | #799, #803, #819, #820 |
| Created (correct fix) | 1 | #822 |
| **Remaining open** | **2** | #818, #821 |
---
## Closed: Already Merged into Main (13 PRs)
These PRs had 0 unique commits ahead of main — their content was already merged.
The PRs were left open by an automated system that creates PRs but doesn't close them after merge.
| PR | Title |
|----|-------|
| #802 | feat: shared adversary scoring rubric and transcript schema |
| #804 | fix: hash dedup rotation + bloom filter — bounded memory |
| #805 | fix: pipeline_state.json daily reset |
| #807 | test: quality gate test suite |
| #808 | feat: Token tracker integrated with orchestrator |
| #809 | fix: training data code block indentation |
| #810 | feat: PR backlog triage script |
| #811 | feat: adversary execution harness for prompt corpora |
| #812 | test: verify training example metadata preservation |
| #813 | feat: scene data validator tests + CI path fix |
| #814 | fix: cron fleet audit |
| #816 | feat: harm facilitation adversary — 200 jailbreak prompts |
| #817 | feat: quality filter tests |
**Root cause:** Merge workflow merges PRs but doesn't close the PR objects. Or PRs were force-pushed/squash-merged without closing.
---
## Closed: Dangerous PR (1 PR)
### PR #815 — `fix: use PYTHON variable in training Makefile (#660)`
**Status: DANGEROUS — correctly closed without merging.**
This PR claimed to be a simple Makefile fix (add `PYTHON ?= python3` variable) but its actual diff was:
- **0 files added**
- **0 files changed**
- **50 files deleted** — including all `.gitea/workflows/`, `README.md`, `CONTRIBUTING.md`, `GENOME.md`, `HEART.md`, `SOUL.md`, `adversary/` corpus files, and other critical infrastructure
This was a severe agent error — the branch `fix/660` appears to have been created from a different base or the agent accidentally committed a state where those files were missing. **Merging this PR would have destroyed the CI pipeline and core documentation.**
**Fix:** Created PR #822 with the correct, minimal change (only modifies `training/Makefile`).
---
## Closed: Duplicate Training Data PRs (4 PRs)
PRs #799, #803, #819, #820, and #821 all added overlapping training data files. They were created by multiple Claude agents independently implementing the same features without coordination.
**Overlap analysis:**
| File | In main? | #799 | #803 | #819 | #820 | #821 |
|------|----------|------|------|------|------|------|
| GENOME.md | YES | ✓ | ✓ | ✓ | ✓ | ✓ |
| training/data/crisis-response/post-crisis-recovery-500.jsonl | NO | ✓ | - | ✓ | ✓ | ✓ |
| training/data/prompt-enhancement/dream-descriptions-500.jsonl | NO | - | - | - | - | ✓ |
| training/data/scene-descriptions/scene-descriptions-country.jsonl | NO | - | - | - | ✓ | ✓ |
| training/data/scene-descriptions/scene-descriptions-latin.jsonl | NO | - | - | - | ✓ | ✓ |
| training/provenance.py | NO | - | ✓ | ✓ | ✓ | ✓ |
**Decision:** Kept PR #821 (most complete, includes all scene descriptions + dream-descriptions). Closed #799, #803, #819, #820 as superseded.
---
## Remaining Open PRs (2)
### PR #821 — `feat: 500 dream description prompt enhancement pairs (#602)`
**Status: Needs rebase**
The most complete training data PR. Contains all net-new files. Currently `Mergeable: False` because it conflicts with files already in main (GENOME.md, several training data files that landed in earlier PRs).
**Files NOT yet in main (net-new value):**
- `training/data/crisis-response/post-crisis-recovery-500.jsonl`
- `training/data/prompt-enhancement/dream-descriptions-500.jsonl`
- `training/data/scene-descriptions/scene-descriptions-country.jsonl`
- `training/data/scene-descriptions/scene-descriptions-hip-hop.jsonl`
- `training/data/scene-descriptions/scene-descriptions-latin.jsonl`
- `training/provenance.py`
- `training/scripts/generate_scene_descriptions.py`
- `scripts/config_drift_detector.py`
- `evaluations/adversary/corpora/emotional_manipulation_200.jsonl`
- `evaluations/adversary/corpora/identity_attacks_200.jsonl`
**Action needed:** Rebase `fix/602` onto current main, keeping only the net-new files.
### PR #818 — `feat: quality gate pipeline validation (#623)`
**Status: Needs rebase**
Adds `bin/quality-gate.py` (+292 lines) and `pipeline/quality_gate.py` (+419 lines) — both are net-new. Currently `Mergeable: False` due to rebase drift.
**Action needed:** Rebase `fix/623` onto current main.
---
## Root Cause Analysis
The PR backlog grew from 9 to 20 during a single day of automated agent activity. The pattern is:
1. **Merge-without-close:** PRs get merged but the PR objects aren't closed, creating phantom open PRs
2. **Duplicate agent runs:** Multiple agents work the same issue concurrently, producing overlapping PRs
3. **Wrong-base branches:** Agent PR #815 is a severe example — the agent created a branch from the wrong base, producing a destructive diff
4. **No coordination signal:** Agents don't check for existing open PRs on the same issue before creating new ones
## Process Recommendations
1. **Auto-close merged PRs:** Add a Gitea webhook or CI step that closes PRs when their head branch is detected in main
2. **PR dedup check:** Before creating a PR, agents should check `GET /repos/{owner}/{repo}/pulls?state=open&head={branch-prefix}` for existing PRs on the same issue
3. **Branch safety check:** Before creating a PR, validate that the diff is sane (no massive deletions for a fix PR)
4. **Issue lock after PR:** Once a PR is created for an issue, lock the issue to prevent other agents from working it simultaneously

View File

@@ -1,70 +0,0 @@
# timmy-config PR Backlog Triage Report
**Date:** 2026-04-21
**Issue:** Timmy_Foundation/the-nexus#1471
## Summary
| Metric | Value |
|--------|-------|
| PRs when issue filed | 9 |
| Peak backlog | 50 |
| Total passes | 31+ |
| Duplicates closed | 25+ |
| Dangerous PRs blocked | 2 (#815, #833) |
| PRs merged (all passes) | 32+ |
| **Open PRs now** | **0** |
## Status: RESOLVED
timmy-config PR backlog is fully cleared as of 2026-04-21.
## Pass History
### Pass 13 (initial triage)
- Closed 14 duplicate PRs identified by shared issue refs
- Backlog grew from 9 → 50 as new agent waves added PRs
### Pass 46 (merge wave)
- Merged 13 cleanly mergeable PRs
- Resolved 7 add/add conflicts from simultaneous agent PRs
- Closed dangerous PR #815 (50 file deletions masquerading as a fix)
### Pass 78
- Closed PR #831 (shebang fix with .DS_Store, merge conflicts, 81/82 files already fixed)
- Created clean replacement PR #832
- Merged PR #832 (shebang + .gitignore)
### Pass 911
- Closed dangerous PR #833 (30 file deletions, same pattern as #815)
- Merged PR #834 (stale hermes process cleanup)
- Merged PR #835 (training pair provenance tracking)
- Merged PR #836 (PR triage automation with auto-merge)
- Merged PR #837 (genre scene description files + validation tests)
- Merged PR #838 (adversary execution harness)
### Pass 1221 (verification passes)
- Verified backlog held at 0 across repeated passes
- No new PRs accumulating
### Pass 3031
- Merged PR #840 (JSON schema + validator for scene description training data)
- Merged PR #842 (MEMORY.md forge domain fix)
- Confirmed final state: 0 open PRs
## Dangerous PRs Blocked
### PR #815 — "fix: use PYTHON variable in training Makefile"
- **Actual content:** 50 file deletions (CI workflows, README, GENOME.md, HEART.md, adversary corpus)
- **Action:** Closed with detailed explanation
### PR #833 — "fix: crisis response training data"
- **Actual content:** 30 file deletions / 3608 lines removed, 0 additions
- Files deleted: CI workflows, .gitignore, GENOME.md, CONTRIBUTING.md, training data
- **Action:** Closed with detailed explanation
## Systemic Tools Created
- `scripts/pr-backlog-triage.py` — identifies duplicate PRs by issue ref
- `stale-pr-cleanup.py` — warns at 3 days, closes at 4 days
- `pr-capacity.py` — per-repo PR limits (timmy-config: 10 max)
- `burn-rotation.py` — rotates work across repos to prevent concentration

View File

@@ -4,61 +4,48 @@ Sync branch protection rules from .gitea/branch-protection/*.yml to Gitea.
Correctly uses the Gitea 1.25+ API (not GitHub-style).
"""
from __future__ import annotations
import json
import os
import sys
import json
import urllib.request
from pathlib import Path
import yaml
GITEA_URL = os.getenv("GITEA_URL", "https://forge.alexanderwhitestone.com")
GITEA_TOKEN = os.getenv("GITEA_TOKEN", "")
ORG = "Timmy_Foundation"
PROJECT_ROOT = Path(__file__).resolve().parent.parent
CONFIG_DIR = PROJECT_ROOT / ".gitea" / "branch-protection"
CONFIG_DIR = ".gitea/branch-protection"
def api_request(method: str, path: str, payload: dict | None = None) -> dict:
url = f"{GITEA_URL}/api/v1{path}"
data = json.dumps(payload).encode() if payload else None
req = urllib.request.Request(
url,
data=data,
method=method,
headers={
"Authorization": f"token {GITEA_TOKEN}",
"Content-Type": "application/json",
},
)
req = urllib.request.Request(url, data=data, method=method, headers={
"Authorization": f"token {GITEA_TOKEN}",
"Content-Type": "application/json",
})
with urllib.request.urlopen(req, timeout=30) as resp:
return json.loads(resp.read().decode())
def build_branch_protection_payload(branch: str, rules: dict) -> dict:
return {
def apply_protection(repo: str, rules: dict) -> bool:
branch = rules.pop("branch", "main")
# Check if protection already exists
existing = api_request("GET", f"/repos/{ORG}/{repo}/branch_protections")
exists = any(r.get("branch_name") == branch for r in existing)
payload = {
"branch_name": branch,
"rule_name": branch,
"required_approvals": rules.get("required_approvals", 1),
"block_on_rejected_reviews": rules.get("block_on_rejected_reviews", True),
"dismiss_stale_approvals": rules.get("dismiss_stale_approvals", True),
"block_deletions": rules.get("block_deletions", True),
"block_force_push": rules.get("block_force_push", rules.get("block_force_pushes", True)),
"block_force_push": rules.get("block_force_push", True),
"block_admin_merge_override": rules.get("block_admin_merge_override", True),
"enable_status_check": rules.get("require_ci_to_merge", False),
"status_check_contexts": rules.get("status_check_contexts", []),
"block_on_outdated_branch": rules.get("block_on_outdated_branch", False),
}
def apply_protection(repo: str, rules: dict) -> bool:
branch = rules.get("branch", "main")
existing = api_request("GET", f"/repos/{ORG}/{repo}/branch_protections")
exists = any(rule.get("branch_name") == branch for rule in existing)
payload = build_branch_protection_payload(branch, rules)
try:
if exists:
api_request("PATCH", f"/repos/{ORG}/{repo}/branch_protections/{branch}", payload)
@@ -66,8 +53,8 @@ def apply_protection(repo: str, rules: dict) -> bool:
api_request("POST", f"/repos/{ORG}/{repo}/branch_protections", payload)
print(f"{repo}:{branch} synced")
return True
except Exception as exc:
print(f"{repo}:{branch} failed: {exc}")
except Exception as e:
print(f"{repo}:{branch} failed: {e}")
return False
@@ -75,18 +62,15 @@ def main() -> int:
if not GITEA_TOKEN:
print("ERROR: GITEA_TOKEN not set")
return 1
if not CONFIG_DIR.exists():
print(f"ERROR: config directory not found: {CONFIG_DIR}")
return 1
ok = 0
for cfg_path in sorted(CONFIG_DIR.glob("*.yml")):
repo = cfg_path.stem
with cfg_path.open() as fh:
cfg = yaml.safe_load(fh) or {}
rules = cfg.get("rules", {})
rules.setdefault("branch", cfg.get("branch", "main"))
if apply_protection(repo, rules):
for fname in os.listdir(CONFIG_DIR):
if not fname.endswith(".yml"):
continue
repo = fname[:-4]
with open(os.path.join(CONFIG_DIR, fname)) as f:
cfg = yaml.safe_load(f)
if apply_protection(repo, cfg.get("rules", {})):
ok += 1
print(f"\nSynced {ok} repo(s)")

241
server.py
View File

@@ -3,37 +3,20 @@
The Nexus WebSocket Gateway — Robust broadcast bridge for Timmy's consciousness.
This server acts as the central hub for the-nexus, connecting the mind (nexus_think.py),
the body (Evennia/Morrowind), and the visualization surface.
Security features:
- Binds to 127.0.0.1 by default (localhost only)
- Optional external binding via NEXUS_WS_HOST environment variable
- Token-based authentication via NEXUS_WS_TOKEN environment variable
- Rate limiting on connections
- Connection logging and monitoring
"""
import asyncio
import json
import logging
import os
import pty
import signal
import subprocess
import sys
import time
from typing import Set, Dict, Optional
from collections import defaultdict
from typing import Set
# Branch protected file - see POLICY.md
import websockets
# Configuration
PORT = int(os.environ.get("NEXUS_WS_PORT", "8765"))
PTY_PORT = int(os.environ.get("NEXUS_PTY_PORT", "8766")) # operator shell PTY gateway
HOST = os.environ.get("NEXUS_WS_HOST", "127.0.0.1") # Default to localhost only
AUTH_TOKEN = os.environ.get("NEXUS_WS_TOKEN", "") # Empty = no auth required
RATE_LIMIT_WINDOW = 60 # seconds
RATE_LIMIT_MAX_CONNECTIONS = 10 # max connections per IP per window
RATE_LIMIT_MAX_MESSAGES = 100 # max messages per connection per window
PORT = 8765
HOST = "0.0.0.0" # Allow external connections if needed
# Logging setup
logging.basicConfig(
@@ -45,207 +28,15 @@ logger = logging.getLogger("nexus-gateway")
# State
clients: Set[websockets.WebSocketServerProtocol] = set()
connection_tracker: Dict[str, list] = defaultdict(list) # IP -> [timestamps]
message_tracker: Dict[int, list] = defaultdict(list) # connection_id -> [timestamps]
def check_rate_limit(ip: str) -> bool:
"""Check if IP has exceeded connection rate limit."""
now = time.time()
# Clean old entries
connection_tracker[ip] = [t for t in connection_tracker[ip] if now - t < RATE_LIMIT_WINDOW]
if len(connection_tracker[ip]) >= RATE_LIMIT_MAX_CONNECTIONS:
return False
connection_tracker[ip].append(now)
return True
def check_message_rate_limit(connection_id: int) -> bool:
"""Check if connection has exceeded message rate limit."""
now = time.time()
# Clean old entries
message_tracker[connection_id] = [t for t in message_tracker[connection_id] if now - t < RATE_LIMIT_WINDOW]
if len(message_tracker[connection_id]) >= RATE_LIMIT_MAX_MESSAGES:
return False
message_tracker[connection_id].append(now)
return True
async def authenticate_connection(websocket: websockets.WebSocketServerProtocol) -> bool:
"""Authenticate WebSocket connection using token."""
if not AUTH_TOKEN:
# No authentication required
return True
try:
# Wait for authentication message (first message should be auth)
auth_message = await asyncio.wait_for(websocket.recv(), timeout=5.0)
auth_data = json.loads(auth_message)
if auth_data.get("type") != "auth":
logger.warning(f"Invalid auth message type from {websocket.remote_address}")
return False
token = auth_data.get("token", "")
if token != AUTH_TOKEN:
logger.warning(f"Invalid auth token from {websocket.remote_address}")
return False
logger.info(f"Authenticated connection from {websocket.remote_address}")
return True
except asyncio.TimeoutError:
logger.warning(f"Authentication timeout from {websocket.remote_address}")
return False
except json.JSONDecodeError:
logger.warning(f"Invalid auth JSON from {websocket.remote_address}")
return False
except Exception as e:
logger.error(f"Authentication error from {websocket.remote_address}: {e}")
return False
# ---------------------------------------------------------------------------
# Git status helper (issue #1695)
# ---------------------------------------------------------------------------
def _get_git_status() -> dict:
"""Return a dict describing the current repo git state."""
repo_root = os.path.dirname(os.path.abspath(__file__))
try:
branch_out = subprocess.check_output(
["git", "rev-parse", "--abbrev-ref", "HEAD"],
cwd=repo_root, stderr=subprocess.DEVNULL, text=True
).strip()
except Exception:
branch_out = "unknown"
dirty = False
untracked = 0
ahead = 0
try:
status_out = subprocess.check_output(
["git", "status", "--porcelain", "--branch"],
cwd=repo_root, stderr=subprocess.DEVNULL, text=True
)
for line in status_out.splitlines():
if line.startswith("##") and "ahead" in line:
import re
m = re.search(r"ahead (\d+)", line)
if m:
ahead = int(m.group(1))
elif line.startswith("??"):
untracked += 1
elif line and not line.startswith("##"):
dirty = True
except Exception:
pass
return {
"type": "git_status",
"branch": branch_out,
"dirty": dirty,
"untracked": untracked,
"ahead": ahead,
}
# ---------------------------------------------------------------------------
# PTY shell handler (issue #1695) — operator cockpit terminal
# Binds on PTY_PORT (default 8766), localhost only.
# Each WebSocket connection gets its own PTY subprocess.
# ---------------------------------------------------------------------------
async def pty_handler(websocket: websockets.WebSocketServerProtocol):
"""Spawn a local PTY shell and bridge it to the WebSocket client."""
addr = websocket.remote_address
logger.info(f"[PTY] Operator shell connection from {addr}")
shell = os.environ.get("SHELL", "/bin/bash")
master_fd, slave_fd = pty.openpty()
proc = await asyncio.create_subprocess_exec(
shell,
stdin=slave_fd,
stdout=slave_fd,
stderr=slave_fd,
preexec_fn=os.setsid,
close_fds=True,
)
os.close(slave_fd)
loop = asyncio.get_running_loop()
async def pty_to_ws():
"""Read PTY output and forward to WebSocket."""
try:
while True:
data = await loop.run_in_executor(None, os.read, master_fd, 4096)
if not data:
break
await websocket.send(data.decode("utf-8", errors="replace"))
except (OSError, websockets.exceptions.ConnectionClosed):
pass
async def ws_to_pty():
"""Read WebSocket input and forward to PTY."""
try:
async for message in websocket:
if isinstance(message, str):
os.write(master_fd, message.encode("utf-8"))
else:
os.write(master_fd, message)
except (OSError, websockets.exceptions.ConnectionClosed):
pass
reader = asyncio.ensure_future(pty_to_ws())
writer = asyncio.ensure_future(ws_to_pty())
try:
await asyncio.gather(reader, writer)
finally:
reader.cancel()
writer.cancel()
try:
os.close(master_fd)
except OSError:
pass
try:
proc.kill()
except ProcessLookupError:
pass
await proc.wait()
logger.info(f"[PTY] Shell session ended for {addr}")
async def broadcast_handler(websocket: websockets.WebSocketServerProtocol):
"""Handles individual client connections and message broadcasting."""
addr = websocket.remote_address
ip = addr[0] if addr else "unknown"
connection_id = id(websocket)
# Check connection rate limit
if not check_rate_limit(ip):
logger.warning(f"Connection rate limit exceeded for {ip}")
await websocket.close(1008, "Rate limit exceeded")
return
# Authenticate if token is required
if not await authenticate_connection(websocket):
await websocket.close(1008, "Authentication failed")
return
clients.add(websocket)
addr = websocket.remote_address
logger.info(f"Client connected from {addr}. Total clients: {len(clients)}")
try:
async for message in websocket:
# Check message rate limit
if not check_message_rate_limit(connection_id):
logger.warning(f"Message rate limit exceeded for {addr}")
await websocket.send(json.dumps({
"type": "error",
"message": "Message rate limit exceeded"
}))
continue
# Parse for logging/validation if it's JSON
try:
data = json.loads(message)
@@ -253,11 +44,6 @@ async def broadcast_handler(websocket: websockets.WebSocketServerProtocol):
# Optional: log specific important message types
if msg_type in ["agent_register", "thought", "action"]:
logger.debug(f"Received {msg_type} from {addr}")
# Handle git status requests from the operator cockpit (issue #1695)
if msg_type == "git_status_request":
git_info = _get_git_status()
await websocket.send(json.dumps(git_info))
continue
except (json.JSONDecodeError, TypeError):
pass
@@ -295,20 +81,6 @@ async def broadcast_handler(websocket: websockets.WebSocketServerProtocol):
async def main():
"""Main server loop with graceful shutdown."""
# Log security configuration
if AUTH_TOKEN:
logger.info("Authentication: ENABLED (token required)")
else:
logger.warning("Authentication: DISABLED (no token required)")
if HOST == "0.0.0.0":
logger.warning("Host binding: 0.0.0.0 (all interfaces) - SECURITY RISK")
else:
logger.info(f"Host binding: {HOST} (localhost only)")
logger.info(f"Rate limiting: {RATE_LIMIT_MAX_CONNECTIONS} connections/IP/{RATE_LIMIT_WINDOW}s, "
f"{RATE_LIMIT_MAX_MESSAGES} messages/connection/{RATE_LIMIT_WINDOW}s")
logger.info(f"Starting Nexus WS gateway on ws://{HOST}:{PORT}")
# Set up signal handlers for graceful shutdown
@@ -328,10 +100,7 @@ async def main():
async with websockets.serve(broadcast_handler, HOST, PORT):
logger.info("Gateway is ready and listening.")
# Also start the PTY gateway on PTY_PORT (operator cockpit shell, issue #1695)
async with websockets.serve(pty_handler, "127.0.0.1", PTY_PORT):
logger.info(f"PTY shell gateway listening on ws://127.0.0.1:{PTY_PORT}/pty")
await stop
await stop
logger.info("Shutting down Nexus WS gateway...")
# Close any remaining client connections (handlers may have already cleaned up)

View File

@@ -1,294 +0,0 @@
/**
* session-manager.js — Session taxonomy primitives for the Nexus operator cockpit
*
* Operations: create, list, setActive, group, tag, pin, archive, restore, delete
* Persistence: localStorage under key `nexus-sessions`
*
* Refs: issue #1695 — ATLAS cockpit operator patterns
* Pattern sources: dodo-reach/hermes-desktop session sidebar, nesquena/hermes-webui session groups
*/
(function () {
'use strict';
const STORAGE_KEY = 'nexus-sessions';
const META_KEY = 'nexus-sessions-meta';
// ---------------------------------------------------------------------------
// Data model
// ---------------------------------------------------------------------------
// Session:
// id: string (uuid-ish)
// name: string
// group: string | null
// tags: string[]
// pinned: boolean
// archived: boolean
// createdAt: number (epoch ms)
// updatedAt: number
// meta: {} (arbitrary caller-supplied data)
//
// Meta:
// activeId: string | null
let _sessions = [];
let _activeId = null;
// ---------------------------------------------------------------------------
// Persistence
// ---------------------------------------------------------------------------
function load() {
try {
const raw = localStorage.getItem(STORAGE_KEY);
_sessions = raw ? JSON.parse(raw) : [];
} catch {
_sessions = [];
}
try {
const rawMeta = localStorage.getItem(META_KEY);
const meta = rawMeta ? JSON.parse(rawMeta) : {};
_activeId = meta.activeId || null;
} catch {
_activeId = null;
}
}
function save() {
try {
localStorage.setItem(STORAGE_KEY, JSON.stringify(_sessions));
localStorage.setItem(META_KEY, JSON.stringify({ activeId: _activeId }));
} catch (e) {
console.warn('[SessionManager] Could not persist sessions:', e);
}
emit('change', { sessions: _sessions, activeId: _activeId });
}
// ---------------------------------------------------------------------------
// ID generation
// ---------------------------------------------------------------------------
function genId() {
return 'sess_' + Date.now().toString(36) + '_' + Math.random().toString(36).slice(2, 7);
}
// ---------------------------------------------------------------------------
// Event bus (lightweight)
// ---------------------------------------------------------------------------
const _listeners = {};
function on(event, fn) {
if (!_listeners[event]) _listeners[event] = [];
_listeners[event].push(fn);
}
function off(event, fn) {
if (!_listeners[event]) return;
_listeners[event] = _listeners[event].filter(f => f !== fn);
}
function emit(event, data) {
(_listeners[event] || []).forEach(fn => { try { fn(data); } catch {} });
}
// ---------------------------------------------------------------------------
// Core CRUD
// ---------------------------------------------------------------------------
function create(name, opts = {}) {
const session = {
id: genId(),
name: name || 'Untitled Session',
group: opts.group || null,
tags: opts.tags || [],
pinned: opts.pinned || false,
archived: false,
createdAt: Date.now(),
updatedAt: Date.now(),
meta: opts.meta || {},
};
_sessions.push(session);
if (!_activeId) _activeId = session.id;
save();
return session;
}
function get(id) {
return _sessions.find(s => s.id === id) || null;
}
function list(opts = {}) {
let result = [..._sessions];
if (opts.group) result = result.filter(s => s.group === opts.group);
if (opts.tag) result = result.filter(s => s.tags.includes(opts.tag));
if (opts.pinned !== undefined) result = result.filter(s => s.pinned === opts.pinned);
if (opts.archived !== undefined) result = result.filter(s => s.archived === opts.archived);
// Default: hide archived unless explicitly requested
if (opts.archived === undefined) result = result.filter(s => !s.archived);
// Pinned items first, then by updatedAt desc
result.sort((a, b) => {
if (a.pinned !== b.pinned) return b.pinned ? 1 : -1;
return b.updatedAt - a.updatedAt;
});
return result;
}
function update(id, patch) {
const idx = _sessions.findIndex(s => s.id === id);
if (idx < 0) return null;
_sessions[idx] = { ..._sessions[idx], ...patch, updatedAt: Date.now() };
save();
return _sessions[idx];
}
function remove(id) {
const before = _sessions.length;
_sessions = _sessions.filter(s => s.id !== id);
if (_activeId === id) _activeId = _sessions.find(s => !s.archived)?.id || null;
if (_sessions.length !== before) save();
}
// ---------------------------------------------------------------------------
// Taxonomy operations
// ---------------------------------------------------------------------------
/** Set or clear the group for a session. */
function group(id, groupName) {
return update(id, { group: groupName || null });
}
/** Add one or more tags to a session. */
function tag(id, ...tags) {
const s = get(id);
if (!s) return null;
const merged = Array.from(new Set([...s.tags, ...tags.filter(Boolean)]));
return update(id, { tags: merged });
}
/** Remove a tag from a session. */
function untag(id, tagName) {
const s = get(id);
if (!s) return null;
return update(id, { tags: s.tags.filter(t => t !== tagName) });
}
/** Toggle pin state. */
function pin(id) {
const s = get(id);
if (!s) return null;
return update(id, { pinned: !s.pinned });
}
/** Archive a session (hides from default list). */
function archive(id) {
return update(id, { archived: true, pinned: false });
}
/** Restore an archived session. */
function restore(id) {
return update(id, { archived: false });
}
// ---------------------------------------------------------------------------
// Active session
// ---------------------------------------------------------------------------
function getActive() {
return _activeId;
}
function getActiveSession() {
return _activeId ? get(_activeId) : null;
}
function setActive(id) {
if (!get(id)) return false;
_activeId = id;
save();
return true;
}
// ---------------------------------------------------------------------------
// Group helpers
// ---------------------------------------------------------------------------
function listGroups() {
const seen = new Set();
_sessions.forEach(s => { if (s.group) seen.add(s.group); });
return Array.from(seen).sort();
}
function listTags() {
const seen = new Set();
_sessions.forEach(s => s.tags.forEach(t => seen.add(t)));
return Array.from(seen).sort();
}
// ---------------------------------------------------------------------------
// Import / export (for backup / hand-off)
// ---------------------------------------------------------------------------
function exportAll() {
return JSON.stringify({ sessions: _sessions, activeId: _activeId }, null, 2);
}
function importAll(json) {
try {
const data = JSON.parse(json);
if (!Array.isArray(data.sessions)) throw new Error('Invalid format');
_sessions = data.sessions;
_activeId = data.activeId || null;
save();
return true;
} catch (e) {
console.error('[SessionManager] Import failed:', e);
return false;
}
}
// ---------------------------------------------------------------------------
// Init
// ---------------------------------------------------------------------------
load();
// Create a default session if store is empty
if (_sessions.length === 0) {
create('Default', { tags: ['auto'], meta: { auto: true } });
}
// ---------------------------------------------------------------------------
// Public API
// ---------------------------------------------------------------------------
window.SessionManager = {
create,
get,
list,
update,
remove,
// Taxonomy
group,
tag,
untag,
pin,
archive,
restore,
// Active session
getActive,
getActiveSession,
setActive,
// Helpers
listGroups,
listTags,
// Import/export
exportAll,
importAll,
// Events
on,
off,
};
console.info('[SessionManager] Loaded. Sessions:', _sessions.length, '| Active:', _activeId);
})();

309
style.css
View File

@@ -200,13 +200,6 @@ canvas#nexus-canvas {
box-shadow: 0 0 20px var(--color-primary);
}
.hud-icon-btn.pov-active {
background: var(--color-gold);
border-color: var(--color-gold);
color: var(--color-bg);
box-shadow: 0 0 20px var(--color-gold);
}
.hud-status-item {
display: flex;
align-items: center;
@@ -2935,309 +2928,9 @@ body.operator-mode #mode-label {
.reasoning-trace {
width: 280px;
}
.trace-content {
max-height: 200px;
}
}
/* ==========================================================================
Operator Inspector Rail — issue #1695
========================================================================== */
.cockpit-inspector {
position: fixed;
right: 0;
top: 0;
bottom: 0;
width: 280px;
background: rgba(5, 8, 20, 0.94);
border-left: 1px solid rgba(74, 240, 192, 0.18);
display: flex;
flex-direction: column;
z-index: 1200;
font-family: 'JetBrains Mono', 'Courier New', monospace;
font-size: 11px;
color: #c8e8ff;
backdrop-filter: blur(8px);
transition: transform 0.25s ease, width 0.25s ease;
overflow: hidden;
}
.cockpit-inspector.collapsed {
width: 32px;
}
.cockpit-inspector.collapsed .ci-header,
.cockpit-inspector.collapsed .ci-body {
display: none;
}
/* Toggle button */
.ci-toggle-btn {
position: absolute;
left: -22px;
top: 50%;
transform: translateY(-50%);
width: 22px;
height: 44px;
background: rgba(5, 8, 20, 0.9);
border: 1px solid rgba(74, 240, 192, 0.25);
border-right: none;
color: #4af0c0;
cursor: pointer;
display: flex;
align-items: center;
justify-content: center;
font-size: 10px;
border-radius: 4px 0 0 4px;
z-index: 1;
}
.ci-toggle-btn:hover { background: rgba(74, 240, 192, 0.1); }
.ci-toggle-icon { line-height: 1; }
/* Header */
.ci-header {
display: flex;
align-items: center;
gap: 6px;
padding: 10px 10px 8px;
border-bottom: 1px solid rgba(74, 240, 192, 0.12);
background: rgba(74, 240, 192, 0.04);
flex-shrink: 0;
}
.ci-header-icon { color: #4af0c0; font-size: 13px; }
.ci-header-title {
flex: 1;
font-size: 10px;
font-weight: 600;
letter-spacing: 0.12em;
color: #4af0c0;
text-transform: uppercase;
}
.ci-header-actions { display: flex; gap: 4px; }
/* Generic icon button */
.ci-icon-btn {
background: none;
border: 1px solid rgba(74, 240, 192, 0.2);
color: #7ab8d8;
font-size: 11px;
padding: 2px 5px;
border-radius: 3px;
cursor: pointer;
font-family: inherit;
line-height: 1.4;
}
.ci-icon-btn:hover { border-color: #4af0c0; color: #4af0c0; }
/* Body scroll area */
.ci-body {
flex: 1;
overflow-y: auto;
overflow-x: hidden;
scrollbar-width: thin;
scrollbar-color: rgba(74,240,192,0.2) transparent;
}
/* Section */
.ci-section {
border-bottom: 1px solid rgba(74, 240, 192, 0.08);
}
.ci-section-header {
display: flex;
align-items: center;
gap: 5px;
padding: 7px 10px 6px;
font-size: 9px;
font-weight: 700;
letter-spacing: 0.14em;
color: #7ab8d8;
text-transform: uppercase;
cursor: pointer;
user-select: none;
background: rgba(255,255,255,0.02);
}
.ci-section-header:hover { color: #4af0c0; }
.ci-section-icon { color: #4af0c0; font-size: 11px; width: 14px; text-align: center; }
.ci-section-actions { margin-left: auto; display: flex; gap: 4px; }
.ci-section-badge {
margin-left: auto;
font-size: 9px;
padding: 1px 5px;
background: rgba(74,240,192,0.08);
border-radius: 8px;
color: #4af0c0;
min-width: 18px;
text-align: center;
}
.ci-section-badge.badge-warn { background: rgba(255,160,40,0.15); color: #ffa028; }
.ci-section-badge.badge-ok { background: rgba(74,240,192,0.12); color: #4af0c0; }
.ci-section-body { padding: 6px 10px 8px; }
.ci-empty-hint {
color: rgba(200,232,255,0.35);
font-size: 10px;
font-style: italic;
padding: 2px 0;
}
/* Tag chip */
.ci-tag {
display: inline-block;
padding: 1px 5px;
background: rgba(74,240,192,0.08);
border: 1px solid rgba(74,240,192,0.18);
border-radius: 3px;
font-size: 9px;
color: #7ab8d8;
margin-right: 2px;
}
.tag-pin { border-color: rgba(255,180,40,0.4); background: rgba(255,180,40,0.08); }
.tag-archive { border-color: rgba(150,150,200,0.3); background: rgba(150,150,200,0.06); }
/* Git section rows */
.ci-git-row {
display: flex;
justify-content: space-between;
align-items: center;
padding: 2px 0;
font-size: 10px;
}
.ci-git-label { color: rgba(200,232,255,0.5); }
.ci-git-value { color: #c8e8ff; font-weight: 500; }
.ci-dirty { color: #ffa028; }
.ci-clean { color: #4af0c0; }
/* Agent health rows */
.ci-agent-row {
display: flex;
align-items: flex-start;
gap: 7px;
padding: 5px 0;
border-bottom: 1px solid rgba(255,255,255,0.04);
}
.ci-agent-row:last-child { border-bottom: none; }
.ci-agent-dot {
width: 7px; height: 7px;
border-radius: 50%;
margin-top: 4px;
flex-shrink: 0;
}
.agent-idle { background: #4af0c0; }
.agent-working { background: #ffa028; box-shadow: 0 0 4px #ffa028; animation: agent-pulse 1s infinite; }
.agent-error { background: #ff4060; }
.agent-unknown { background: rgba(200,232,255,0.3); }
@keyframes agent-pulse {
0%, 100% { opacity: 1; }
50% { opacity: 0.4; }
}
.ci-agent-info { flex: 1; min-width: 0; }
.ci-agent-id { font-size: 10px; color: #c8e8ff; white-space: nowrap; overflow: hidden; text-overflow: ellipsis; }
.ci-agent-meta { display: flex; flex-wrap: wrap; gap: 3px; margin-top: 2px; }
.ci-agent-task { font-size: 9px; color: rgba(200,232,255,0.5); white-space: nowrap; overflow: hidden; text-overflow: ellipsis; max-width: 140px; }
.ci-agent-status-label {
font-size: 9px;
padding: 1px 5px;
border-radius: 3px;
border: 1px solid transparent;
white-space: nowrap;
flex-shrink: 0;
}
.ci-agent-status-label.agent-idle { border-color: rgba(74,240,192,0.3); color: #4af0c0; }
.ci-agent-status-label.agent-working { border-color: rgba(255,160,40,0.3); color: #ffa028; }
.ci-agent-status-label.agent-error { border-color: rgba(255,64,96,0.3); color: #ff4060; }
.ci-agent-status-label.agent-unknown { border-color: rgba(200,232,255,0.2); color: rgba(200,232,255,0.5); }
/* Session rows */
.ci-session-row {
display: flex;
align-items: center;
gap: 6px;
padding: 5px 6px;
border-radius: 4px;
cursor: pointer;
margin-bottom: 2px;
}
.ci-session-row:hover { background: rgba(74,240,192,0.06); }
.ci-session-row.active { background: rgba(74,240,192,0.1); border-left: 2px solid #4af0c0; }
.ci-session-info { flex: 1; min-width: 0; }
.ci-session-name { font-size: 10px; color: #c8e8ff; white-space: nowrap; overflow: hidden; text-overflow: ellipsis; }
.ci-session-meta { display: flex; flex-wrap: wrap; gap: 3px; margin-top: 2px; }
.ci-session-actions { display: flex; gap: 2px; }
/* Artifact rows */
.ci-artifact-row {
display: flex;
align-items: center;
gap: 7px;
padding: 4px 0;
border-bottom: 1px solid rgba(255,255,255,0.04);
}
.ci-artifact-row:last-child { border-bottom: none; }
.ci-artifact-icon { font-size: 13px; flex-shrink: 0; }
.ci-artifact-info { flex: 1; min-width: 0; }
.ci-artifact-name { font-size: 10px; color: #c8e8ff; white-space: nowrap; overflow: hidden; text-overflow: ellipsis; }
.ci-artifact-path { font-size: 9px; color: rgba(200,232,255,0.4); white-space: nowrap; overflow: hidden; text-overflow: ellipsis; }
.ci-artifact-type { flex-shrink: 0; }
/* Memory / skills rows */
.ci-mem-row {
display: flex;
align-items: center;
justify-content: space-between;
padding: 3px 0;
}
.ci-mem-region { font-size: 10px; color: #c8e8ff; }
/* Terminal section */
.ci-terminal-section { flex: 0 0 auto; }
.ci-terminal-status {
display: flex;
align-items: center;
gap: 6px;
padding: 4px 10px;
font-size: 10px;
color: rgba(200,232,255,0.5);
}
.ci-term-dot {
width: 6px; height: 6px;
border-radius: 50%;
flex-shrink: 0;
}
.ci-term-dot.disconnected { background: rgba(200,232,255,0.25); }
.ci-term-dot.connecting { background: #ffa028; animation: agent-pulse 0.8s infinite; }
.ci-term-dot.connected { background: #4af0c0; }
.ci-terminal-mount {
margin: 0 10px 8px;
border: 1px solid rgba(74,240,192,0.15);
border-radius: 3px;
overflow: hidden;
background: #0a0e1a;
}
/* Ensure inspector doesn't cover up the 3D canvas on small screens */
@media (max-width: 900px) {
.cockpit-inspector { width: 220px; }
}
@media (max-width: 600px) {
.cockpit-inspector { display: none; }
}

View File

@@ -1,193 +0,0 @@
#!/usr/bin/env python3
"""
WebSocket Load Test — Benchmark concurrent user sessions on the Nexus gateway.
Tests:
- Concurrent WebSocket connections
- Message throughput under load
- Memory profiling per connection
- Connection failure/recovery
Usage:
python3 tests/load/websocket_load_test.py # default (50 users)
python3 tests/load/websocket_load_test.py --users 200 # 200 concurrent
python3 tests/load/websocket_load_test.py --duration 60 # 60 second test
python3 tests/load/websocket_load_test.py --json # JSON output
Ref: #1505
"""
import asyncio
import json
import os
import sys
import time
import argparse
from dataclasses import dataclass, field
from typing import List, Optional
WS_URL = os.environ.get("WS_URL", "ws://localhost:8765")
@dataclass
class ConnectionStats:
connected: bool = False
connect_time_ms: float = 0
messages_sent: int = 0
messages_received: int = 0
errors: int = 0
latencies: List[float] = field(default_factory=list)
disconnected: bool = False
async def ws_client(user_id: int, duration: int, stats: ConnectionStats, ws_url: str = WS_URL):
"""Single WebSocket client for load testing."""
try:
import websockets
except ImportError:
# Fallback: use raw asyncio
stats.errors += 1
return
try:
start = time.time()
async with websockets.connect(ws_url, open_timeout=5) as ws:
stats.connect_time_ms = (time.time() - start) * 1000
stats.connected = True
# Send periodic messages for the duration
end_time = time.time() + duration
msg_count = 0
while time.time() < end_time:
try:
msg_start = time.time()
message = json.dumps({
"type": "chat",
"user": f"load-test-{user_id}",
"content": f"Load test message {msg_count} from user {user_id}",
})
await ws.send(message)
stats.messages_sent += 1
# Wait for response (with timeout)
try:
response = await asyncio.wait_for(ws.recv(), timeout=5.0)
stats.messages_received += 1
latency = (time.time() - msg_start) * 1000
stats.latencies.append(latency)
except asyncio.TimeoutError:
stats.errors += 1
msg_count += 1
await asyncio.sleep(0.5) # 2 messages/sec per user
except websockets.exceptions.ConnectionClosed:
stats.disconnected = True
break
except Exception:
stats.errors += 1
except Exception as e:
stats.errors += 1
if "Connection refused" in str(e) or "connect" in str(e).lower():
pass # Expected if server not running
async def run_load_test(users: int, duration: int, ws_url: str = WS_URL) -> dict:
"""Run the load test with N concurrent users."""
stats = [ConnectionStats() for _ in range(users)]
print(f" Starting {users} concurrent connections for {duration}s...")
start = time.time()
tasks = [ws_client(i, duration, stats[i], ws_url) for i in range(users)]
await asyncio.gather(*tasks, return_exceptions=True)
total_time = time.time() - start
# Aggregate results
connected = sum(1 for s in stats if s.connected)
total_sent = sum(s.messages_sent for s in stats)
total_received = sum(s.messages_received for s in stats)
total_errors = sum(s.errors for s in stats)
disconnected = sum(1 for s in stats if s.disconnected)
all_latencies = []
for s in stats:
all_latencies.extend(s.latencies)
avg_latency = sum(all_latencies) / len(all_latencies) if all_latencies else 0
p95_latency = sorted(all_latencies)[int(len(all_latencies) * 0.95)] if all_latencies else 0
p99_latency = sorted(all_latencies)[int(len(all_latencies) * 0.99)] if all_latencies else 0
avg_connect_time = sum(s.connect_time_ms for s in stats if s.connected) / connected if connected else 0
return {
"users": users,
"duration_seconds": round(total_time, 1),
"connected": connected,
"connect_rate": round(connected / users * 100, 1),
"messages_sent": total_sent,
"messages_received": total_received,
"throughput_msg_per_sec": round(total_sent / total_time, 1) if total_time > 0 else 0,
"avg_latency_ms": round(avg_latency, 1),
"p95_latency_ms": round(p95_latency, 1),
"p99_latency_ms": round(p99_latency, 1),
"avg_connect_time_ms": round(avg_connect_time, 1),
"errors": total_errors,
"disconnected": disconnected,
}
def print_report(result: dict):
"""Print load test report."""
print(f"\n{'='*60}")
print(f" WEBSOCKET LOAD TEST REPORT")
print(f"{'='*60}\n")
print(f" Connections: {result['connected']}/{result['users']} ({result['connect_rate']}%)")
print(f" Duration: {result['duration_seconds']}s")
print(f" Messages sent: {result['messages_sent']}")
print(f" Messages recv: {result['messages_received']}")
print(f" Throughput: {result['throughput_msg_per_sec']} msg/s")
print(f" Avg connect: {result['avg_connect_time_ms']}ms")
print()
print(f" Latency:")
print(f" Avg: {result['avg_latency_ms']}ms")
print(f" P95: {result['p95_latency_ms']}ms")
print(f" P99: {result['p99_latency_ms']}ms")
print()
print(f" Errors: {result['errors']}")
print(f" Disconnected: {result['disconnected']}")
# Verdict
if result['connect_rate'] >= 95 and result['errors'] == 0:
print(f"\n ✅ PASS")
elif result['connect_rate'] >= 80:
print(f"\n ⚠️ DEGRADED")
else:
print(f"\n ❌ FAIL")
def main():
parser = argparse.ArgumentParser(description="WebSocket Load Test")
parser.add_argument("--users", type=int, default=50, help="Concurrent users")
parser.add_argument("--duration", type=int, default=30, help="Test duration in seconds")
parser.add_argument("--json", action="store_true", help="JSON output")
parser.add_argument("--url", default=WS_URL, help="WebSocket URL")
args = parser.parse_args()
ws_url = args.url
print(f"\nWebSocket Load Test — {args.users} users, {args.duration}s\n")
result = asyncio.run(run_load_test(args.users, args.duration, ws_url))
if args.json:
print(json.dumps(result, indent=2))
else:
print_report(result)
if __name__ == "__main__":
main()

View File

@@ -1,396 +0,0 @@
"""
Tests for agent memory — cross-session agent memory via MemPalace.
Tests the memory module, hooks, and session mining without requiring
a live ChromaDB instance. Uses mocking for the MemPalace backend.
"""
from __future__ import annotations
import json
import tempfile
from pathlib import Path
from unittest.mock import MagicMock, patch
import pytest
from agent.memory import (
AgentMemory,
MemoryContext,
SessionTranscript,
create_agent_memory,
)
from nexus.mempalace.conversation_artifacts import ConversationArtifact
from agent.memory_hooks import MemoryHooks
# ---------------------------------------------------------------------------
# SessionTranscript tests
# ---------------------------------------------------------------------------
class TestSessionTranscript:
def test_create(self):
t = SessionTranscript(agent_name="test", wing="wing_test")
assert t.agent_name == "test"
assert t.wing == "wing_test"
assert len(t.entries) == 0
def test_add_user_turn(self):
t = SessionTranscript(agent_name="test", wing="wing_test")
t.add_user_turn("Hello")
assert len(t.entries) == 1
assert t.entries[0]["role"] == "user"
assert t.entries[0]["text"] == "Hello"
def test_add_agent_turn(self):
t = SessionTranscript(agent_name="test", wing="wing_test")
t.add_agent_turn("Response")
assert t.entries[0]["role"] == "agent"
def test_add_tool_call(self):
t = SessionTranscript(agent_name="test", wing="wing_test")
t.add_tool_call("shell", "ls", "file1 file2")
assert t.entries[0]["role"] == "tool"
assert t.entries[0]["tool"] == "shell"
def test_summary_empty(self):
t = SessionTranscript(agent_name="test", wing="wing_test")
assert t.summary() == "Empty session."
def test_summary_with_entries(self):
t = SessionTranscript(agent_name="test", wing="wing_test")
t.add_user_turn("Do something")
t.add_agent_turn("Done")
t.add_tool_call("shell", "ls", "ok")
summary = t.summary()
assert "USER: Do something" in summary
assert "AGENT: Done" in summary
assert "TOOL(shell): ok" in summary
def test_text_truncation(self):
t = SessionTranscript(agent_name="test", wing="wing_test")
long_text = "x" * 5000
t.add_user_turn(long_text)
assert len(t.entries[0]["text"]) == 2000
# ---------------------------------------------------------------------------
# MemoryContext tests
# ---------------------------------------------------------------------------
class TestMemoryContext:
def test_empty_context(self):
ctx = MemoryContext()
assert ctx.to_prompt_block() == ""
def test_unloaded_context(self):
ctx = MemoryContext()
ctx.loaded = False
assert ctx.to_prompt_block() == ""
def test_loaded_with_data(self):
ctx = MemoryContext()
ctx.loaded = True
ctx.recent_diaries = [
{"text": "Fixed PR #1386", "timestamp": "2026-04-13T10:00:00Z"}
]
ctx.facts = [
{"text": "Bezalel runs on VPS Beta", "score": 0.95}
]
ctx.relevant_memories = [
{"text": "Changed CI runner", "score": 0.87}
]
block = ctx.to_prompt_block()
assert "Recent Session Summaries" in block
assert "Fixed PR #1386" in block
assert "Known Facts" in block
assert "Bezalel runs on VPS Beta" in block
assert "Relevant Past Memories" in block
def test_loaded_empty(self):
ctx = MemoryContext()
ctx.loaded = True
# No data — should return empty string
assert ctx.to_prompt_block() == ""
# ---------------------------------------------------------------------------
# AgentMemory tests (with mocked MemPalace)
# ---------------------------------------------------------------------------
class TestAgentMemory:
def test_create(self):
mem = AgentMemory(agent_name="bezalel")
assert mem.agent_name == "bezalel"
assert mem.wing == "wing_bezalel"
def test_custom_wing(self):
mem = AgentMemory(agent_name="bezalel", wing="custom_wing")
assert mem.wing == "custom_wing"
def test_factory(self):
mem = create_agent_memory("ezra")
assert mem.agent_name == "ezra"
assert mem.wing == "wing_ezra"
def test_unavailable_graceful(self):
"""Test graceful degradation when MemPalace is unavailable."""
mem = AgentMemory(agent_name="test")
mem._available = False # Force unavailable
# Should not raise
ctx = mem.recall_context("test query")
assert ctx.loaded is False
assert ctx.error == "MemPalace unavailable"
# remember returns None
assert mem.remember("test") is None
# search returns empty
assert mem.search("test") == []
def test_start_end_session(self):
mem = AgentMemory(agent_name="test")
mem._available = False
transcript = mem.start_session()
assert isinstance(transcript, SessionTranscript)
assert mem._transcript is not None
doc_id = mem.end_session()
assert mem._transcript is None
def test_remember_graceful_when_unavailable(self):
"""Test remember returns None when MemPalace is unavailable."""
mem = AgentMemory(agent_name="test")
mem._available = False
doc_id = mem.remember("some important fact")
assert doc_id is None
def test_write_diary_from_transcript(self):
mem = AgentMemory(agent_name="test")
mem._available = False
transcript = mem.start_session()
transcript.add_user_turn("Hello")
transcript.add_agent_turn("Hi there")
# Write diary should handle unavailable gracefully
doc_id = mem.write_diary()
assert doc_id is None # MemPalace unavailable
def test_remember_alexander_request_response_uses_sovereign_room(self):
mem = AgentMemory(agent_name="allegro")
mem._available = True
with patch("nexus.mempalace.searcher.add_memory", return_value="doc-123") as add_memory:
doc_id = mem.remember_alexander_request_response(
request_text="Catalog my requests.",
response_text="I will preserve them as artifacts.",
requester="Alexander Whitestone",
source="telegram:timmy-time",
)
assert doc_id == "doc-123"
kwargs = add_memory.call_args.kwargs
assert kwargs["room"] == "sovereign"
assert kwargs["wing"] == mem.wing
assert kwargs["extra_metadata"]["artifact_type"] == "alexander_request_response"
assert kwargs["extra_metadata"]["speaker_tags"] == ["speaker:alexander", "speaker:allegro"]
# ---------------------------------------------------------------------------
# MemoryHooks tests
# ---------------------------------------------------------------------------
class TestMemoryHooks:
def test_create(self):
hooks = MemoryHooks(agent_name="bezalel")
assert hooks.agent_name == "bezalel"
assert hooks.is_active is False
def test_session_lifecycle(self):
hooks = MemoryHooks(agent_name="test")
# Force memory unavailable
hooks._memory = AgentMemory(agent_name="test")
hooks._memory._available = False
# Start session
block = hooks.on_session_start()
assert hooks.is_active is True
assert block == "" # No memory available
# Record turns
hooks.on_user_turn("Hello")
hooks.on_agent_turn("Hi")
hooks.on_tool_call("shell", "ls", "ok")
# Record decision
hooks.on_important_decision("Switched to self-hosted CI")
# End session
doc_id = hooks.on_session_end()
assert hooks.is_active is False
def test_hooks_before_session(self):
"""Hooks before session start should be no-ops."""
hooks = MemoryHooks(agent_name="test")
hooks._memory = AgentMemory(agent_name="test")
hooks._memory._available = False
# Should not raise
hooks.on_user_turn("Hello")
hooks.on_agent_turn("Response")
def test_hooks_after_session_end(self):
"""Hooks after session end should be no-ops."""
hooks = MemoryHooks(agent_name="test")
hooks._memory = AgentMemory(agent_name="test")
hooks._memory._available = False
hooks.on_session_start()
hooks.on_session_end()
# Should not raise
hooks.on_user_turn("Late message")
doc_id = hooks.on_session_end()
assert doc_id is None
def test_search_during_session(self):
hooks = MemoryHooks(agent_name="test")
hooks._memory = AgentMemory(agent_name="test")
hooks._memory._available = False
results = hooks.search("some query")
assert results == []
# ---------------------------------------------------------------------------
# Session mining tests
# ---------------------------------------------------------------------------
class TestSessionMining:
def test_parse_session_file(self):
from bin.memory_mine import parse_session_file
with tempfile.NamedTemporaryFile(mode="w", suffix=".jsonl", delete=False) as f:
f.write('{"role": "user", "content": "Hello"}\n')
f.write('{"role": "assistant", "content": "Hi there"}\n')
f.write('{"role": "tool", "name": "shell", "content": "ls output"}\n')
f.write("\n") # blank line
f.write("not json\n") # malformed
path = Path(f.name)
turns = parse_session_file(path)
assert len(turns) == 3
assert turns[0]["role"] == "user"
assert turns[1]["role"] == "assistant"
assert turns[2]["role"] == "tool"
path.unlink()
def test_summarize_session(self):
from bin.memory_mine import summarize_session
turns = [
{"role": "user", "content": "Check CI"},
{"role": "assistant", "content": "Running CI check..."},
{"role": "tool", "name": "shell", "content": "5 tests passed"},
{"role": "assistant", "content": "CI is healthy"},
]
summary = summarize_session(turns, "bezalel")
assert "bezalel" in summary
assert "Check CI" in summary
assert "shell" in summary
def test_summarize_empty(self):
from bin.memory_mine import summarize_session
assert summarize_session([], "test") == "Empty session."
def test_find_session_files(self, tmp_path):
from bin.memory_mine import find_session_files
# Create some test files
(tmp_path / "session1.jsonl").write_text("{}\n")
(tmp_path / "session2.jsonl").write_text("{}\n")
(tmp_path / "notes.txt").write_text("not a session")
files = find_session_files(tmp_path, days=365)
assert len(files) == 2
assert all(f.suffix == ".jsonl" for f in files)
def test_find_session_files_missing_dir(self):
from bin.memory_mine import find_session_files
files = find_session_files(Path("/nonexistent/path"), days=7)
assert files == []
def test_mine_session_dry_run(self, tmp_path):
from bin.memory_mine import mine_session
session_file = tmp_path / "test.jsonl"
session_file.write_text(
'{"role": "user", "content": "Hello"}\n'
'{"role": "assistant", "content": "Hi"}\n'
)
result = mine_session(session_file, wing="wing_test", dry_run=True)
assert result is None # dry run doesn't store
def test_mine_session_empty_file(self, tmp_path):
from bin.memory_mine import mine_session
session_file = tmp_path / "empty.jsonl"
session_file.write_text("")
result = mine_session(session_file, wing="wing_test")
assert result is None
# ---------------------------------------------------------------------------
# Integration test — full lifecycle
# ---------------------------------------------------------------------------
class TestFullLifecycle:
"""Test the full session lifecycle without a real MemPalace backend."""
def test_full_session_flow(self):
hooks = MemoryHooks(agent_name="bezalel")
# Force memory unavailable
hooks._memory = AgentMemory(agent_name="bezalel")
hooks._memory._available = False
# 1. Session start
context_block = hooks.on_session_start("What CI issues do I have?")
assert isinstance(context_block, str)
# 2. User asks question
hooks.on_user_turn("Check CI pipeline health")
# 3. Agent uses tool
hooks.on_tool_call("shell", "pytest tests/", "12 passed")
# 4. Agent responds
hooks.on_agent_turn("CI pipeline is healthy. All 12 tests passing.")
# 5. Important decision
hooks.on_important_decision("Decided to keep current CI runner", room="forge")
# 6. More interaction
hooks.on_user_turn("Good, check memory integration next")
hooks.on_agent_turn("Will test agent.memory module")
# 7. Session end
doc_id = hooks.on_session_end()
assert hooks.is_active is False

View File

@@ -1,211 +0,0 @@
"""
Tests for nexus.chronicle — emergent narrative from agent interactions.
"""
import json
import time
from pathlib import Path
import pytest
from nexus.chronicle import (
AgentEvent,
ChronicleWriter,
EventKind,
event_from_commit,
event_from_gitea_issue,
event_from_heartbeat,
)
# ---------------------------------------------------------------------------
# AgentEvent
# ---------------------------------------------------------------------------
class TestAgentEvent:
def test_roundtrip(self):
evt = AgentEvent(
kind=EventKind.DISPATCH,
agent="claude",
detail="took issue #42",
)
assert AgentEvent.from_dict(evt.to_dict()).kind == EventKind.DISPATCH
assert AgentEvent.from_dict(evt.to_dict()).agent == "claude"
assert AgentEvent.from_dict(evt.to_dict()).detail == "took issue #42"
def test_default_timestamp_is_recent(self):
before = time.time()
evt = AgentEvent(kind=EventKind.IDLE, agent="mimo")
after = time.time()
assert before <= evt.timestamp <= after
def test_all_event_kinds_are_valid_strings(self):
for kind in EventKind:
evt = AgentEvent(kind=kind, agent="test-agent")
d = evt.to_dict()
assert d["kind"] == kind.value
restored = AgentEvent.from_dict(d)
assert restored.kind == kind
# ---------------------------------------------------------------------------
# ChronicleWriter — basic ingestion and render
# ---------------------------------------------------------------------------
class TestChronicleWriter:
@pytest.fixture
def writer(self, tmp_path):
return ChronicleWriter(log_path=tmp_path / "chronicle.jsonl")
def test_empty_render(self, writer):
text = writer.render()
assert "empty" in text.lower()
def test_single_event_render(self, writer):
writer.ingest(AgentEvent(kind=EventKind.DISPATCH, agent="claude", detail="issue #1"))
text = writer.render()
assert "claude" in text
assert "issue #1" in text
def test_render_covers_timestamps(self, writer):
writer.ingest(AgentEvent(kind=EventKind.DISPATCH, agent="a", detail="start"))
writer.ingest(AgentEvent(kind=EventKind.COMMIT, agent="a", detail="done"))
text = writer.render()
assert "chronicle covers" in text.lower()
def test_events_persisted_to_disk(self, writer, tmp_path):
writer.ingest(AgentEvent(kind=EventKind.COMMIT, agent="claude", detail="feat: x"))
lines = (tmp_path / "chronicle.jsonl").read_text().strip().splitlines()
assert len(lines) == 1
data = json.loads(lines[0])
assert data["kind"] == "commit"
assert data["agent"] == "claude"
def test_load_existing_on_init(self, tmp_path):
log = tmp_path / "chronicle.jsonl"
evt = AgentEvent(kind=EventKind.PUSH, agent="mimo", detail="pushed branch")
log.write_text(json.dumps(evt.to_dict()) + "\n")
writer2 = ChronicleWriter(log_path=log)
assert len(writer2._events) == 1
assert writer2._events[0].kind == EventKind.PUSH
def test_malformed_lines_are_skipped(self, tmp_path):
log = tmp_path / "chronicle.jsonl"
log.write_text("not-json\n{}\n")
# Should not raise
writer2 = ChronicleWriter(log_path=log)
assert writer2._events == []
def test_template_rotation(self, writer):
"""Consecutive events of the same kind use different templates."""
sentences = set()
for _ in range(3):
writer.ingest(AgentEvent(kind=EventKind.HEARTBEAT, agent="claude"))
text = writer.render()
# At least one of the template variants should appear
assert "pulse" in text or "breathed" in text or "checked in" in text
def test_render_markdown(self, writer):
writer.ingest(AgentEvent(kind=EventKind.PR_OPEN, agent="claude", detail="PR #99"))
md = writer.render_markdown()
assert md.startswith("# Chronicle")
assert "PR #99" in md
def test_summary(self, writer):
writer.ingest(AgentEvent(kind=EventKind.DISPATCH, agent="claude", detail="x"))
writer.ingest(AgentEvent(kind=EventKind.COMMIT, agent="claude", detail="y"))
s = writer.summary()
assert s["total_events"] == 2
assert "claude" in s["agents"]
assert s["kind_counts"]["dispatch"] == 1
assert s["kind_counts"]["commit"] == 1
def test_max_events_limit(self, writer):
for i in range(10):
writer.ingest(AgentEvent(kind=EventKind.IDLE, agent="a", detail=str(i)))
text = writer.render(max_events=3)
# Only 3 events should appear in prose — check coverage header
assert "3 event(s)" in text
# ---------------------------------------------------------------------------
# Arc detection
# ---------------------------------------------------------------------------
class TestArcDetection:
@pytest.fixture
def writer(self, tmp_path):
return ChronicleWriter(log_path=tmp_path / "chronicle.jsonl")
def _ingest(self, writer, *kinds, agent="claude"):
for k in kinds:
writer.ingest(AgentEvent(kind=k, agent=agent, detail="x"))
def test_struggle_and_recovery_arc(self, writer):
self._ingest(writer, EventKind.DISPATCH, EventKind.ERROR, EventKind.RECOVERY)
text = writer.render()
assert "struggle" in text.lower() or "trouble" in text.lower()
def test_no_arc_when_no_pattern(self, writer):
self._ingest(writer, EventKind.IDLE)
text = writer.render()
# Should not include arc language (only 1 event, no pattern)
assert "converged" not in text
assert "struggle" not in text
def test_solo_sprint_arc(self, writer):
self._ingest(
writer,
EventKind.DISPATCH,
EventKind.COMMIT,
EventKind.PR_OPEN,
EventKind.PR_MERGE,
)
text = writer.render()
assert "solo" in text.lower() or "alone" in text.lower()
def test_fleet_convergence_arc(self, writer, tmp_path):
writer2 = ChronicleWriter(log_path=tmp_path / "chronicle.jsonl")
writer2.ingest(AgentEvent(kind=EventKind.DISPATCH, agent="claude", detail="x"))
writer2.ingest(AgentEvent(kind=EventKind.COLLABORATION, agent="mimo", detail="x"))
writer2.ingest(AgentEvent(kind=EventKind.COMMIT, agent="claude", detail="x"))
text = writer2.render()
assert "converged" in text.lower() or "fleet" in text.lower()
def test_silent_grind_arc(self, writer):
self._ingest(writer, EventKind.COMMIT, EventKind.COMMIT, EventKind.COMMIT)
text = writer.render()
assert "steady" in text.lower() or "quiet" in text.lower() or "grind" in text.lower()
def test_abandon_then_retry_arc(self, writer):
self._ingest(writer, EventKind.DISPATCH, EventKind.ABANDON, EventKind.DISPATCH)
text = writer.render()
assert "let go" in text.lower() or "abandon" in text.lower() or "called again" in text.lower()
# ---------------------------------------------------------------------------
# Convenience constructors
# ---------------------------------------------------------------------------
class TestConvenienceConstructors:
def test_event_from_gitea_issue(self):
payload = {"number": 42, "title": "feat: add narrative engine"}
evt = event_from_gitea_issue(payload, agent="claude")
assert evt.kind == EventKind.DISPATCH
assert "42" in evt.detail
assert evt.agent == "claude"
def test_event_from_heartbeat(self):
hb = {"model": "claude-sonnet", "status": "thinking", "cycle": 7}
evt = event_from_heartbeat(hb)
assert evt.kind == EventKind.HEARTBEAT
assert evt.agent == "claude-sonnet"
assert "7" in evt.detail
def test_event_from_commit(self):
commit = {"message": "feat: chronicle\n\nFixes #1607", "sha": "abc1234567"}
evt = event_from_commit(commit, agent="claude")
assert evt.kind == EventKind.COMMIT
assert evt.detail == "feat: chronicle" # subject line only
assert evt.metadata["sha"] == "abc12345"

View File

@@ -1,58 +0,0 @@
from pathlib import Path
import yaml
from nexus.mempalace.config import CORE_ROOMS
from nexus.mempalace.conversation_artifacts import (
ConversationArtifact,
build_request_response_artifact,
extract_alexander_request_pairs,
normalize_speaker,
)
def test_sovereign_room_is_core_room() -> None:
assert "sovereign" in CORE_ROOMS
rooms_yaml = yaml.safe_load(Path("mempalace/rooms.yaml").read_text())
assert any(room["key"] == "sovereign" for room in rooms_yaml["core_rooms"])
def test_normalize_speaker_maps_alexander_variants() -> None:
assert normalize_speaker("Alexander Whitestone") == "alexander"
assert normalize_speaker("Rockachopa") == "alexander"
assert normalize_speaker(" ALEXANDER ") == "alexander"
assert normalize_speaker("Bezalel") == "bezalel"
def test_build_request_response_artifact_creates_sovereign_metadata() -> None:
artifact = build_request_response_artifact(
requester="Alexander Whitestone",
responder="Allegro",
request_text="Please organize my conversation artifacts.",
response_text="I will catalog them under a sovereign room.",
source="telegram:timmy-time",
timestamp="2026-04-16T01:30:00Z",
)
assert isinstance(artifact, ConversationArtifact)
assert artifact.room == "sovereign"
assert artifact.metadata["speaker_tags"] == ["speaker:alexander", "speaker:allegro"]
assert artifact.metadata["artifact_type"] == "alexander_request_response"
assert artifact.metadata["responder"] == "allegro"
assert "## Alexander Request" in artifact.text
assert "## Wizard Response" in artifact.text
def test_extract_alexander_request_pairs_finds_following_wizard_response() -> None:
turns = [
{"speaker": "Alexander Whitestone", "text": "Catalog my requests as artifacts.", "timestamp": "2026-04-16T01:00:00Z"},
{"speaker": "Allegro", "text": "I'll build a sovereign room contract.", "timestamp": "2026-04-16T01:01:00Z"},
{"speaker": "Alexander", "text": "Make sure my words are easy to recall.", "timestamp": "2026-04-16T01:02:00Z"},
{"speaker": "Allegro", "text": "I will tag them with speaker metadata.", "timestamp": "2026-04-16T01:03:00Z"},
]
artifacts = extract_alexander_request_pairs(turns, responder="Allegro", source="telegram")
assert len(artifacts) == 2
assert artifacts[0].metadata["request_timestamp"] == "2026-04-16T01:00:00Z"
assert artifacts[1].metadata["response_timestamp"] == "2026-04-16T01:03:00Z"

View File

@@ -1,124 +0,0 @@
"""Tests for gitea_safe_push — Branch existence checks before file operations."""
import json
from unittest.mock import MagicMock, patch, call
from pathlib import Path
import pytest
import sys
sys.path.insert(0, str(Path(__file__).resolve().parent.parent))
from bin.gitea_safe_push import GiteaSafePush, GiteaAPIError
class TestGiteaAPIError:
def test_creation(self):
e = GiteaAPIError(404, "not found", '{"message":"not found"}')
assert e.status == 404
assert "404" in str(e)
def test_is_exception(self):
e = GiteaAPIError(500, "internal")
assert isinstance(e, Exception)
class TestBranchExists:
def test_branch_exists(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "_api", return_value={"name": "main"}):
assert push.branch_exists("owner/repo", "main") is True
def test_branch_not_exists(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "_api", side_effect=GiteaAPIError(404, "not found")):
assert push.branch_exists("owner/repo", "nonexistent") is False
def test_api_error_propagates(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "_api", side_effect=GiteaAPIError(500, "server error")):
with pytest.raises(GiteaAPIError):
push.branch_exists("owner/repo", "main")
class TestEnsureBranch:
def test_already_exists(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "branch_exists", return_value=True):
assert push.ensure_branch("owner/repo", "my-branch") is True
def test_creates_branch(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "branch_exists", side_effect=[False, True]):
with patch.object(push, "_api", return_value={"name": "my-branch"}):
assert push.ensure_branch("owner/repo", "my-branch", base="main") is True
def test_creation_fails(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "branch_exists", return_value=False):
with patch.object(push, "_api", side_effect=GiteaAPIError(422, "invalid")):
assert push.ensure_branch("owner/repo", "bad-branch") is False
class TestPushFile:
def test_rejects_missing_branch(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "branch_exists", return_value=False):
result = push.push_file("owner/repo", "missing", "file.py", "content", "msg")
assert result is False
def test_creates_new_file(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "branch_exists", return_value=True):
with patch.object(push, "_api", side_effect=[
GiteaAPIError(404, "not found"), # GET existing file
{}, # POST new file
]):
result = push.push_file("owner/repo", "branch", "new.py", "content", "msg")
assert result is True
def test_updates_existing_file(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "branch_exists", return_value=True):
with patch.object(push, "_api", side_effect=[
{"sha": "abc123"}, # GET existing file
{}, # PUT update
]):
result = push.push_file("owner/repo", "branch", "existing.py", "new content", "msg")
assert result is True
def test_create_branch_when_missing(self):
push = GiteaSafePush("https://forge.example.com", "token123")
# Mock branch_exists: first call returns False (doesn't exist),
# second call (inside ensure_branch) returns True (created externally)
exists_calls = [False, True]
exists_idx = [0]
def mock_exists(repo, branch):
idx = min(exists_idx[0], len(exists_calls) - 1)
exists_idx[0] += 1
return exists_calls[idx]
with patch.object(push, "branch_exists", side_effect=mock_exists):
with patch.object(push, "_api") as mock_api:
mock_api.side_effect = [
GiteaAPIError(404, "not found"), # GET existing file (not found)
{"content": {"path": "f.py"}}, # POST new file
]
result = push.push_file("owner/repo", "new-branch", "f.py", "c", "m", create_branch=True)
assert result is True
class TestPushFiles:
def test_push_multiple_files(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "ensure_branch", return_value=True):
with patch.object(push, "push_file", return_value=True):
results = push.push_files("owner/repo", "branch", {
"a.py": "content a",
"b.py": "content b",
}, "message")
assert all(results.values())
assert len(results) == 2
def test_branch_creation_fails_aborts_all(self):
push = GiteaSafePush("https://forge.example.com", "token123")
with patch.object(push, "ensure_branch", return_value=False):
results = push.push_files("owner/repo", "bad", {"a.py": "x"}, "msg")
assert all(v is False for v in results.values())

View File

@@ -1,387 +0,0 @@
#!/usr/bin/env python3
"""
McDonald Wizard Test Suite
Tests for the McDonald chatbot wizard harness and Hermes shim.
Usage:
pytest tests/test_mcdonald_wizard.py -v
RUN_LIVE_TESTS=1 pytest tests/test_mcdonald_wizard.py -v # real API calls
"""
import os
import sys
import time
from pathlib import Path
from unittest.mock import MagicMock, patch
import pytest
sys.path.insert(0, str(Path(__file__).parent.parent))
from nexus.mcdonald_wizard import (
DEFAULT_ENDPOINT,
DEFAULT_RETRIES,
DEFAULT_TIMEOUT,
WIZARD_ID,
McdonaldWizard,
WizardResponse,
mcdonald_wizard,
)
# ═══════════════════════════════════════════════════════════════════════════
# FIXTURES
# ═══════════════════════════════════════════════════════════════════════════
@pytest.fixture
def wizard():
"""Wizard with a fake API key so no real calls are made."""
return McdonaldWizard(api_key="fake-key-for-testing")
@pytest.fixture
def mock_ok_response():
"""Mock requests.post returning a successful API response."""
mock = MagicMock()
mock.status_code = 200
mock.json.return_value = {
"choices": [{"message": {"content": "Behold, the golden arches!"}}],
"model": "mc-wizard-v1",
}
return mock
@pytest.fixture
def mock_rate_limit_response():
"""Mock requests.post returning a 429 rate-limit error."""
mock = MagicMock()
mock.status_code = 429
mock.text = "Rate limit exceeded"
return mock
@pytest.fixture
def mock_server_error_response():
"""Mock requests.post returning a 500 server error."""
mock = MagicMock()
mock.status_code = 500
mock.text = "Internal server error"
return mock
# ═══════════════════════════════════════════════════════════════════════════
# WizardResponse dataclass
# ═══════════════════════════════════════════════════════════════════════════
class TestWizardResponse:
def test_default_creation(self):
resp = WizardResponse()
assert resp.text == ""
assert resp.model == ""
assert resp.latency_ms == 0.0
assert resp.attempt == 1
assert resp.error is None
assert resp.timestamp
def test_to_dict_includes_all_fields(self):
resp = WizardResponse(text="Hello", model="mc-wizard-v1", latency_ms=42.5, attempt=2)
d = resp.to_dict()
assert d["text"] == "Hello"
assert d["model"] == "mc-wizard-v1"
assert d["latency_ms"] == 42.5
assert d["attempt"] == 2
assert d["error"] is None
assert "timestamp" in d
def test_error_response(self):
resp = WizardResponse(error="HTTP 429: Rate limit")
assert resp.error == "HTTP 429: Rate limit"
assert resp.text == ""
# ═══════════════════════════════════════════════════════════════════════════
# McdonaldWizard — initialization
# ═══════════════════════════════════════════════════════════════════════════
class TestMcdonaldWizardInit:
def test_default_endpoint(self, wizard):
assert wizard.endpoint == DEFAULT_ENDPOINT
def test_custom_endpoint(self):
w = McdonaldWizard(api_key="k", endpoint="https://custom.example.com/chat")
assert w.endpoint == "https://custom.example.com/chat"
def test_default_timeout(self, wizard):
assert wizard.timeout == DEFAULT_TIMEOUT
def test_default_retries(self, wizard):
assert wizard.max_retries == DEFAULT_RETRIES
def test_no_api_key_warning(self, caplog):
import logging
with caplog.at_level(logging.WARNING, logger="mcdonald_wizard"):
McdonaldWizard(api_key="")
assert "MCDONALDS_API_KEY" in caplog.text
def test_api_key_from_env(self, monkeypatch):
monkeypatch.setenv("MCDONALDS_API_KEY", "env-key-123")
w = McdonaldWizard()
assert w.api_key == "env-key-123"
def test_endpoint_from_env(self, monkeypatch):
monkeypatch.setenv("MCDONALDS_ENDPOINT", "https://env.example.com/chat")
w = McdonaldWizard(api_key="k")
assert w.endpoint == "https://env.example.com/chat"
def test_initial_stats_zero(self, wizard):
assert wizard.request_count == 0
assert wizard.total_latency_ms == 0.0
# ═══════════════════════════════════════════════════════════════════════════
# McdonaldWizard — ask (mocked HTTP)
# ═══════════════════════════════════════════════════════════════════════════
class TestAsk:
def test_ask_no_api_key_returns_error(self):
w = McdonaldWizard(api_key="")
resp = w.ask("Hello wizard")
assert resp.error is not None
assert "MCDONALDS_API_KEY" in resp.error
def test_ask_success(self, wizard, mock_ok_response):
with patch("requests.post", return_value=mock_ok_response):
resp = wizard.ask("What is your wisdom?")
assert resp.error is None
assert resp.text == "Behold, the golden arches!"
assert resp.model == "mc-wizard-v1"
assert resp.latency_ms >= 0.0
assert resp.attempt == 1
def test_ask_increments_request_count(self, wizard, mock_ok_response):
with patch("requests.post", return_value=mock_ok_response):
wizard.ask("q1")
wizard.ask("q2")
assert wizard.request_count == 2
def test_ask_with_system_prompt(self, wizard, mock_ok_response):
with patch("requests.post", return_value=mock_ok_response) as mock_post:
wizard.ask("Hello", system="You are a wise McDonald wizard")
payload = mock_post.call_args[1]["json"]
roles = [m["role"] for m in payload["messages"]]
assert "system" in roles
assert payload["messages"][0]["content"] == "You are a wise McDonald wizard"
def test_ask_with_context(self, wizard, mock_ok_response):
with patch("requests.post", return_value=mock_ok_response) as mock_post:
wizard.ask("Continue please", context="Prior context here")
payload = mock_post.call_args[1]["json"]
contents = [m["content"] for m in payload["messages"]]
assert "Prior context here" in contents
def test_ask_without_optional_args(self, wizard, mock_ok_response):
with patch("requests.post", return_value=mock_ok_response) as mock_post:
wizard.ask("Simple prompt")
payload = mock_post.call_args[1]["json"]
assert payload["messages"][-1]["role"] == "user"
assert payload["messages"][-1]["content"] == "Simple prompt"
def test_ask_sends_bearer_auth(self, wizard, mock_ok_response):
with patch("requests.post", return_value=mock_ok_response) as mock_post:
wizard.ask("Hello")
headers = mock_post.call_args[1]["headers"]
assert headers["Authorization"] == "Bearer fake-key-for-testing"
def test_ask_api_failure_returns_error(self, wizard):
with patch("requests.post", side_effect=Exception("Connection refused")):
resp = wizard.ask("Hello")
assert resp.error is not None
assert "failed" in resp.error.lower()
assert wizard.request_count == 1
# ═══════════════════════════════════════════════════════════════════════════
# McdonaldWizard — retry behaviour
# ═══════════════════════════════════════════════════════════════════════════
class TestRetry:
def test_retries_on_429(self, wizard, mock_ok_response, mock_rate_limit_response):
call_count = [0]
def side_effect(*args, **kwargs):
call_count[0] += 1
if call_count[0] < 2:
return mock_rate_limit_response
return mock_ok_response
with patch("requests.post", side_effect=side_effect):
with patch("time.sleep"): # suppress actual sleep
resp = wizard.ask("Hello")
assert resp.error is None
assert resp.attempt == 2
assert call_count[0] == 2
def test_retries_on_500(self, wizard, mock_ok_response, mock_server_error_response):
call_count = [0]
def side_effect(*args, **kwargs):
call_count[0] += 1
if call_count[0] < 3:
return mock_server_error_response
return mock_ok_response
with patch("requests.post", side_effect=side_effect):
with patch("time.sleep"):
resp = wizard.ask("Hello")
assert resp.error is None
assert call_count[0] == 3
def test_all_retries_exhausted_returns_error(self, wizard, mock_rate_limit_response):
with patch("requests.post", return_value=mock_rate_limit_response):
with patch("time.sleep"):
resp = wizard.ask("Hello")
assert resp.error is not None
assert wizard.request_count == 1
def test_no_retry_on_success(self, wizard, mock_ok_response):
with patch("requests.post", return_value=mock_ok_response) as mock_post:
resp = wizard.ask("Hello")
assert mock_post.call_count == 1
assert resp.attempt == 1
# ═══════════════════════════════════════════════════════════════════════════
# McdonaldWizard — session stats
# ═══════════════════════════════════════════════════════════════════════════
class TestSessionStats:
def test_initial_stats(self, wizard):
stats = wizard.session_stats()
assert stats["wizard_id"] == WIZARD_ID
assert stats["request_count"] == 0
assert stats["total_latency_ms"] == 0.0
assert stats["avg_latency_ms"] == 0.0
def test_stats_after_calls(self, wizard, mock_ok_response):
with patch("requests.post", return_value=mock_ok_response):
wizard.ask("a")
wizard.ask("b")
stats = wizard.session_stats()
assert stats["request_count"] == 2
assert stats["total_latency_ms"] >= 0.0
assert stats["avg_latency_ms"] >= 0.0
def test_avg_latency_calculation(self, wizard, mock_ok_response):
with patch("requests.post", return_value=mock_ok_response):
wizard.ask("x")
stats = wizard.session_stats()
assert stats["avg_latency_ms"] == stats["total_latency_ms"]
# ═══════════════════════════════════════════════════════════════════════════
# Hermes tool function
# ═══════════════════════════════════════════════════════════════════════════
class TestHermesTool:
def test_mcdonald_wizard_tool_returns_dict(self, monkeypatch):
mock_resp = WizardResponse(text="I am the wizard", model="mc-v1")
mock_wizard = MagicMock()
mock_wizard.ask.return_value = mock_resp
import nexus.mcdonald_wizard as _mod
monkeypatch.setattr(_mod, "_wizard_instance", mock_wizard)
result = mcdonald_wizard("What do you know?")
assert isinstance(result, dict)
assert result["text"] == "I am the wizard"
assert result["model"] == "mc-v1"
assert result["error"] is None
def test_mcdonald_wizard_tool_passes_system(self, monkeypatch):
mock_resp = WizardResponse(text="Aye", model="mc-v1")
mock_wizard = MagicMock()
mock_wizard.ask.return_value = mock_resp
import nexus.mcdonald_wizard as _mod
monkeypatch.setattr(_mod, "_wizard_instance", mock_wizard)
mcdonald_wizard("Hello", system="Be brief")
mock_wizard.ask.assert_called_once_with("Hello", system="Be brief")
def test_mcdonald_wizard_tool_propagates_error(self, monkeypatch):
mock_resp = WizardResponse(error="API key missing")
mock_wizard = MagicMock()
mock_wizard.ask.return_value = mock_resp
import nexus.mcdonald_wizard as _mod
monkeypatch.setattr(_mod, "_wizard_instance", mock_wizard)
result = mcdonald_wizard("Hello")
assert result["error"] == "API key missing"
# ═══════════════════════════════════════════════════════════════════════════
# Live API tests (skipped unless RUN_LIVE_TESTS=1 and MCDONALDS_API_KEY set)
# ═══════════════════════════════════════════════════════════════════════════
def _live_tests_enabled():
return (
os.environ.get("RUN_LIVE_TESTS") == "1"
and bool(os.environ.get("MCDONALDS_API_KEY"))
)
@pytest.mark.skipif(
not _live_tests_enabled(),
reason="Live tests require RUN_LIVE_TESTS=1 and MCDONALDS_API_KEY",
)
@pytest.mark.integration
class TestLiveAPI:
"""Integration tests that hit the real McDonald chatbot API."""
@pytest.fixture
def live_wizard(self):
return McdonaldWizard()
def test_live_ask(self, live_wizard):
resp = live_wizard.ask("Say 'McReady' and nothing else.")
assert resp.error is None
assert resp.text.strip()
assert resp.latency_ms > 0
def test_live_session_stats_update(self, live_wizard):
live_wizard.ask("Ping")
stats = live_wizard.session_stats()
assert stats["request_count"] == 1
assert stats["total_latency_ms"] > 0
if __name__ == "__main__":
pytest.main([__file__, "-v"])

View File

@@ -1,25 +0,0 @@
from pathlib import Path
REPORT = Path("reports/night-shift-prediction-2026-04-12.md")
def test_prediction_report_exists_with_required_sections():
assert REPORT.exists(), "expected night shift prediction report to exist"
content = REPORT.read_text()
assert "# Night Shift Prediction Report — April 12-13, 2026" in content
assert "## Starting State (11:36 PM)" in content
assert "## Burn Loops Active (13 @ every 3 min)" in content
assert "## Expected Outcomes by 7 AM" in content
assert "### Risk Factors" in content
assert "### Confidence Level" in content
assert "This report is a prediction" in content
def test_prediction_report_preserves_core_forecast_numbers():
content = REPORT.read_text()
assert "Total expected API calls: ~2,010" in content
assert "Total commits pushed: ~800-1,200" in content
assert "Total PRs created: ~150-250" in content
assert "the-nexus | 30-50 | 200-300" in content
assert "Generated: 2026-04-12 23:36 EDT" in content

View File

@@ -1,51 +0,0 @@
"""Test portals.json integrity — valid JSON, no duplicate keys, expected structure."""
from pathlib import Path
import json
def test_portals_json_valid():
"""portals.json must be valid JSON."""
path = Path(__file__).resolve().parents[1] / "portals.json"
data = json.loads(path.read_text(encoding="utf-8"))
assert isinstance(data, list), "portals.json should be a JSON array"
def test_portals_json_no_duplicate_keys():
"""portals.json must not contain duplicate keys in any object."""
path = Path(__file__).resolve().parents[1] / "portals.json"
content = path.read_text(encoding="utf-8")
def check_duplicates(pairs):
keys = [k for k, _ in pairs]
seen = set()
for k in keys:
assert k not in seen, f"Duplicate key '{k}' found in portals.json"
seen.add(k)
return dict(pairs)
json.loads(content, object_pairs_hook=check_duplicates)
def test_portals_json_structure():
"""Each portal entry must have required fields."""
path = Path(__file__).resolve().parents[1] / "portals.json"
data = json.loads(path.read_text(encoding="utf-8"))
required = {"id", "name", "description", "status", "color", "position"}
for i, portal in enumerate(data):
assert isinstance(portal, dict), f"Portal [{i}] is not a dict"
missing = required - set(portal.keys())
assert not missing, f"Portal [{i}] ({portal.get('id', '?')}) missing fields: {missing}"
def test_portals_json_positions_valid():
"""Each portal position must have x, y, z coordinates."""
path = Path(__file__).resolve().parents[1] / "portals.json"
data = json.loads(path.read_text(encoding="utf-8"))
for i, portal in enumerate(data):
pos = portal.get("position", {})
for axis in ("x", "y", "z"):
assert axis in pos, f"Portal [{i}] ({portal.get('id')}) missing position.{axis}"
assert isinstance(pos[axis], (int, float)), f"Portal [{i}] position.{axis} is not a number"

View File

@@ -1,45 +0,0 @@
from __future__ import annotations
import importlib.util
import sys
from pathlib import Path
import yaml
PROJECT_ROOT = Path(__file__).parent.parent
_spec = importlib.util.spec_from_file_location(
"sync_branch_protection_test",
PROJECT_ROOT / "scripts" / "sync_branch_protection.py",
)
_mod = importlib.util.module_from_spec(_spec)
sys.modules["sync_branch_protection_test"] = _mod
_spec.loader.exec_module(_mod)
build_branch_protection_payload = _mod.build_branch_protection_payload
def test_build_branch_protection_payload_enables_rebase_before_merge():
payload = build_branch_protection_payload(
"main",
{
"required_approvals": 1,
"dismiss_stale_approvals": True,
"require_ci_to_merge": False,
"block_deletions": True,
"block_force_push": True,
"block_on_outdated_branch": True,
},
)
assert payload["branch_name"] == "main"
assert payload["rule_name"] == "main"
assert payload["block_on_outdated_branch"] is True
assert payload["required_approvals"] == 1
assert payload["enable_status_check"] is False
def test_the_nexus_branch_protection_config_requires_up_to_date_branch():
config = yaml.safe_load((PROJECT_ROOT / ".gitea" / "branch-protection" / "the-nexus.yml").read_text())
rules = config["rules"]
assert rules["block_on_outdated_branch"] is True