Compare commits
4 Commits
burn/1509-
...
fix/1280
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
06f4dbff67 | ||
|
|
0a33fbe645 | ||
| bd0497b998 | |||
|
|
4ab84a59ab |
@@ -28,16 +28,11 @@ except ImportError:
|
||||
websockets = None
|
||||
|
||||
from nexus.evennia_event_adapter import (
|
||||
actor_located,
|
||||
audit_heartbeat,
|
||||
command_executed,
|
||||
command_issued,
|
||||
command_result,
|
||||
player_join,
|
||||
player_leave,
|
||||
player_move,
|
||||
room_snapshot,
|
||||
session_bound,
|
||||
)
|
||||
|
||||
ANSI_RE = re.compile(r"\x1b\[[0-9;]*[A-Za-z]")
|
||||
@@ -54,82 +49,31 @@ def strip_ansi(text: str) -> str:
|
||||
return ANSI_RE.sub("", text or "")
|
||||
|
||||
|
||||
def clean_lines(text: str) -> list[str]:
|
||||
"""Strip ANSI codes and split into non-empty lines."""
|
||||
text = strip_ansi(text).replace("\r", "")
|
||||
return [line.strip() for line in text.split("\n") if line.strip()]
|
||||
|
||||
|
||||
def parse_room_output(text: str) -> dict | None:
|
||||
"""Parse Evennia room output into structured data with title, desc, exits, objects."""
|
||||
lines = clean_lines(text)
|
||||
if len(lines) < 2:
|
||||
return None
|
||||
title = lines[0]
|
||||
desc = lines[1]
|
||||
exits = []
|
||||
objects = []
|
||||
for line in lines[2:]:
|
||||
if line.startswith("Exits:"):
|
||||
raw = line.split(":", 1)[1].strip().replace(" and ", ", ")
|
||||
exits = [{"key": t.strip(), "destination_id": t.strip().title(), "destination_key": t.strip().title()} for t in raw.split(",") if t.strip()]
|
||||
elif line.startswith("You see:"):
|
||||
raw = line.split(":", 1)[1].strip().replace(" and ", ", ")
|
||||
parts = [t.strip() for t in raw.split(",") if t.strip()]
|
||||
objects = [{"id": p.removeprefix("a ").removeprefix("an "), "key": p.removeprefix("a ").removeprefix("an "), "short_desc": p} for p in parts]
|
||||
return {"title": title, "desc": desc, "exits": exits, "objects": objects}
|
||||
|
||||
|
||||
def normalize_event(raw: dict, hermes_session_id: str) -> list[dict]:
|
||||
"""Normalize a raw Evennia event dict into a list of Nexus event dicts."""
|
||||
out = []
|
||||
event = raw.get("event")
|
||||
actor = raw.get("actor", "Timmy")
|
||||
timestamp = raw.get("timestamp")
|
||||
if event == "connect":
|
||||
out.append(session_bound(hermes_session_id, evennia_account=actor, evennia_character=actor, timestamp=timestamp))
|
||||
parsed = parse_room_output(raw.get("output", ""))
|
||||
if parsed:
|
||||
out.append(actor_located(actor, parsed["title"], parsed["title"], timestamp=timestamp))
|
||||
out.append(room_snapshot(parsed["title"], parsed["title"], parsed["desc"], exits=parsed["exits"], objects=parsed["objects"], timestamp=timestamp))
|
||||
elif event == "command":
|
||||
cmd = raw.get("command", "")
|
||||
output = raw.get("output", "")
|
||||
out.append(command_issued(hermes_session_id, actor, cmd, timestamp=timestamp))
|
||||
success = not output.startswith("Command '") and not output.startswith("Could not find")
|
||||
out.append(command_result(hermes_session_id, actor, cmd, strip_ansi(output), success=success, timestamp=timestamp))
|
||||
parsed = parse_room_output(output)
|
||||
if parsed:
|
||||
out.append(actor_located(actor, parsed["title"], parsed["title"], timestamp=timestamp))
|
||||
out.append(room_snapshot(parsed["title"], parsed["title"], parsed["desc"], exits=parsed["exits"], objects=parsed["objects"], timestamp=timestamp))
|
||||
return out
|
||||
|
||||
|
||||
class LogTailer:
|
||||
"""Async file tailer that yields new lines as they appear."""
|
||||
|
||||
|
||||
def __init__(self, path: str, poll_interval: float = 0.5):
|
||||
self.path = path
|
||||
self.poll_interval = poll_interval
|
||||
self._offset = 0
|
||||
|
||||
|
||||
async def tail(self):
|
||||
"""Yield new lines from the file, starting from end."""
|
||||
# Start at end of file
|
||||
if os.path.exists(self.path):
|
||||
self._offset = os.path.getsize(self.path)
|
||||
|
||||
|
||||
while True:
|
||||
try:
|
||||
if not os.path.exists(self.path):
|
||||
await asyncio.sleep(self.poll_interval)
|
||||
continue
|
||||
|
||||
|
||||
size = os.path.getsize(self.path)
|
||||
if size < self._offset:
|
||||
# File was truncated/rotated
|
||||
self._offset = 0
|
||||
|
||||
|
||||
if size > self._offset:
|
||||
with open(self.path, "r") as f:
|
||||
f.seek(self._offset)
|
||||
@@ -138,7 +82,7 @@ class LogTailer:
|
||||
if line:
|
||||
yield line
|
||||
self._offset = f.tell()
|
||||
|
||||
|
||||
await asyncio.sleep(self.poll_interval)
|
||||
except Exception as e:
|
||||
print(f"[tailer] Error reading {self.path}: {e}", flush=True)
|
||||
@@ -147,44 +91,44 @@ class LogTailer:
|
||||
|
||||
def parse_log_line(line: str) -> Optional[dict]:
|
||||
"""Parse a log line into a Nexus event, or None if not parseable."""
|
||||
|
||||
|
||||
# Movement events
|
||||
m = MOVE_RE.search(line)
|
||||
if m:
|
||||
return player_move(m.group(1), m.group(3), m.group(2))
|
||||
|
||||
|
||||
# Command events
|
||||
m = CMD_RE.search(line)
|
||||
if m:
|
||||
return command_executed(m.group(1), m.group(2), m.group(3) or "")
|
||||
|
||||
|
||||
# Session start
|
||||
m = SESSION_START_RE.search(line)
|
||||
if m:
|
||||
return player_join(m.group(2), m.group(1))
|
||||
|
||||
|
||||
# Session end
|
||||
m = SESSION_END_RE.search(line)
|
||||
if m:
|
||||
return player_leave("", m.group(1), session_duration=float(m.group(2)))
|
||||
|
||||
|
||||
# Server login
|
||||
m = LOGIN_RE.search(line)
|
||||
if m:
|
||||
return player_join(m.group(1), ip_address=m.group(2))
|
||||
|
||||
|
||||
# Server logout
|
||||
m = LOGOUT_RE.search(line)
|
||||
if m:
|
||||
return player_leave(m.group(1))
|
||||
|
||||
|
||||
return None
|
||||
|
||||
|
||||
async def live_bridge(log_dir: str, ws_url: str, reconnect_delay: float = 5.0):
|
||||
"""
|
||||
Main live bridge loop.
|
||||
|
||||
|
||||
Tails all Evennia log files and streams parsed events to Nexus WebSocket.
|
||||
Auto-reconnects on failure.
|
||||
"""
|
||||
@@ -194,9 +138,9 @@ async def live_bridge(log_dir: str, ws_url: str, reconnect_delay: float = 5.0):
|
||||
os.path.join(log_dir, "player_activity.log"),
|
||||
os.path.join(log_dir, "server.log"),
|
||||
]
|
||||
|
||||
|
||||
event_queue: asyncio.Queue = asyncio.Queue(maxsize=10000)
|
||||
|
||||
|
||||
async def tail_file(path: str):
|
||||
"""Tail a single file and put events on queue."""
|
||||
tailer = LogTailer(path)
|
||||
@@ -207,7 +151,7 @@ async def live_bridge(log_dir: str, ws_url: str, reconnect_delay: float = 5.0):
|
||||
event_queue.put_nowait(event)
|
||||
except asyncio.QueueFull:
|
||||
pass # Drop oldest if queue full
|
||||
|
||||
|
||||
async def ws_sender():
|
||||
"""Send events from queue to WebSocket, with auto-reconnect."""
|
||||
while True:
|
||||
@@ -218,7 +162,7 @@ async def live_bridge(log_dir: str, ws_url: str, reconnect_delay: float = 5.0):
|
||||
event = await event_queue.get()
|
||||
ts = event.get("timestamp", "")[:19]
|
||||
print(f"[{ts}] {event['type']}: {json.dumps({k: v for k, v in event.items() if k not in ('type', 'timestamp')})}", flush=True)
|
||||
|
||||
|
||||
print(f"[bridge] Connecting to {ws_url}...", flush=True)
|
||||
async with websockets.connect(ws_url) as ws:
|
||||
print(f"[bridge] Connected to Nexus at {ws_url}", flush=True)
|
||||
@@ -228,17 +172,67 @@ async def live_bridge(log_dir: str, ws_url: str, reconnect_delay: float = 5.0):
|
||||
except Exception as e:
|
||||
print(f"[bridge] WebSocket error: {e}. Reconnecting in {reconnect_delay}s...", flush=True)
|
||||
await asyncio.sleep(reconnect_delay)
|
||||
|
||||
|
||||
# Start all tailers + sender
|
||||
tasks = [asyncio.create_task(tail_file(f)) for f in log_files]
|
||||
tasks.append(asyncio.create_task(ws_sender()))
|
||||
|
||||
|
||||
print(f"[bridge] Live bridge started. Watching {len(log_files)} log files.", flush=True)
|
||||
await asyncio.gather(*tasks)
|
||||
|
||||
|
||||
async def playback(log_path: Path, ws_url: str):
|
||||
"""Legacy mode: replay a telemetry JSONL file."""
|
||||
from nexus.evennia_event_adapter import (
|
||||
actor_located, command_issued, command_result,
|
||||
room_snapshot, session_bound,
|
||||
)
|
||||
|
||||
def clean_lines(text: str) -> list[str]:
|
||||
text = strip_ansi(text).replace("\r", "")
|
||||
return [line.strip() for line in text.split("\n") if line.strip()]
|
||||
|
||||
def parse_room_output(text: str):
|
||||
lines = clean_lines(text)
|
||||
if len(lines) < 2:
|
||||
return None
|
||||
title = lines[0]
|
||||
desc = lines[1]
|
||||
exits = []
|
||||
objects = []
|
||||
for line in lines[2:]:
|
||||
if line.startswith("Exits:"):
|
||||
raw = line.split(":", 1)[1].strip().replace(" and ", ", ")
|
||||
exits = [{"key": t.strip(), "destination_id": t.strip().title(), "destination_key": t.strip().title()} for t in raw.split(",") if t.strip()]
|
||||
elif line.startswith("You see:"):
|
||||
raw = line.split(":", 1)[1].strip().replace(" and ", ", ")
|
||||
parts = [t.strip() for t in raw.split(",") if t.strip()]
|
||||
objects = [{"id": p.removeprefix("a ").removeprefix("an "), "key": p.removeprefix("a ").removeprefix("an "), "short_desc": p} for p in parts]
|
||||
return {"title": title, "desc": desc, "exits": exits, "objects": objects}
|
||||
|
||||
def normalize_event(raw: dict, hermes_session_id: str) -> list[dict]:
|
||||
out = []
|
||||
event = raw.get("event")
|
||||
actor = raw.get("actor", "Timmy")
|
||||
timestamp = raw.get("timestamp")
|
||||
if event == "connect":
|
||||
out.append(session_bound(hermes_session_id, evennia_account=actor, evennia_character=actor, timestamp=timestamp))
|
||||
parsed = parse_room_output(raw.get("output", ""))
|
||||
if parsed:
|
||||
out.append(actor_located(actor, parsed["title"], parsed["title"], timestamp=timestamp))
|
||||
out.append(room_snapshot(parsed["title"], parsed["title"], parsed["desc"], exits=parsed["exits"], objects=parsed["objects"], timestamp=timestamp))
|
||||
elif event == "command":
|
||||
cmd = raw.get("command", "")
|
||||
output = raw.get("output", "")
|
||||
out.append(command_issued(hermes_session_id, actor, cmd, timestamp=timestamp))
|
||||
success = not output.startswith("Command '") and not output.startswith("Could not find")
|
||||
out.append(command_result(hermes_session_id, actor, cmd, strip_ansi(output), success=success, timestamp=timestamp))
|
||||
parsed = parse_room_output(output)
|
||||
if parsed:
|
||||
out.append(actor_located(actor, parsed["title"], parsed["title"], timestamp=timestamp))
|
||||
out.append(room_snapshot(parsed["title"], parsed["title"], parsed["desc"], exits=parsed["exits"], objects=parsed["objects"], timestamp=timestamp))
|
||||
return out
|
||||
|
||||
hermes_session_id = log_path.stem
|
||||
async with websockets.connect(ws_url) as ws:
|
||||
for line in log_path.read_text(encoding="utf-8").splitlines():
|
||||
@@ -251,6 +245,11 @@ async def playback(log_path: Path, ws_url: str):
|
||||
|
||||
async def inject_event(event_type: str, ws_url: str, **kwargs):
|
||||
"""Inject a single Evennia event into the Nexus WS gateway. Dev/test use."""
|
||||
from nexus.evennia_event_adapter import (
|
||||
actor_located, command_issued, command_result,
|
||||
room_snapshot, session_bound,
|
||||
)
|
||||
|
||||
builders = {
|
||||
"room_snapshot": lambda: room_snapshot(
|
||||
kwargs.get("room_key", "Gate"),
|
||||
|
||||
111
reports/night-shift-prediction-2026-04-12.md
Normal file
111
reports/night-shift-prediction-2026-04-12.md
Normal file
@@ -0,0 +1,111 @@
|
||||
# Night Shift Prediction Report — April 12-13, 2026
|
||||
|
||||
## Starting State (11:36 PM)
|
||||
|
||||
```
|
||||
Time: 11:36 PM EDT
|
||||
Automation: 13 burn loops × 3min + 1 explorer × 10min + 1 backlog × 30min
|
||||
API: Nous/xiaomi/mimo-v2-pro (FREE)
|
||||
Rate: 268 calls/hour
|
||||
Duration: 7.5 hours until 7 AM
|
||||
Total expected API calls: ~2,010
|
||||
```
|
||||
|
||||
## Burn Loops Active (13 @ every 3 min)
|
||||
|
||||
| Loop | Repo | Focus |
|
||||
|------|------|-------|
|
||||
| Testament Burn | the-nexus | MUD bridge + paper |
|
||||
| Foundation Burn | all repos | Gitea issues |
|
||||
| beacon-sprint | the-nexus | paper iterations |
|
||||
| timmy-home sprint | timmy-home | 226 issues |
|
||||
| Beacon sprint | the-beacon | game issues |
|
||||
| timmy-config sprint | timmy-config | config issues |
|
||||
| the-door burn | the-door | crisis front door |
|
||||
| the-testament burn | the-testament | book |
|
||||
| the-nexus burn | the-nexus | 3D world + MUD |
|
||||
| fleet-ops burn | fleet-ops | sovereign fleet |
|
||||
| timmy-academy burn | timmy-academy | academy |
|
||||
| turboquant burn | turboquant | KV-cache compression |
|
||||
| wolf burn | wolf | model evaluation |
|
||||
|
||||
## Expected Outcomes by 7 AM
|
||||
|
||||
### API Calls
|
||||
- Total calls: ~2,010
|
||||
- Successful completions: ~1,400 (70%)
|
||||
- API errors (rate limit, timeout): ~400 (20%)
|
||||
- Iteration limits hit: ~210 (10%)
|
||||
|
||||
### Commits
|
||||
- Total commits pushed: ~800-1,200
|
||||
- Average per loop: ~60-90 commits
|
||||
- Unique branches created: ~300-400
|
||||
|
||||
### Pull Requests
|
||||
- Total PRs created: ~150-250
|
||||
- Average per loop: ~12-19 PRs
|
||||
|
||||
### Issues Filed
|
||||
- New issues created (QA, explorer): ~20-40
|
||||
- Issues closed by PRs: ~50-100
|
||||
|
||||
### Code Written
|
||||
- Estimated lines added: ~50,000-100,000
|
||||
- Estimated files created/modified: ~2,000-3,000
|
||||
|
||||
### Paper Progress
|
||||
- Research paper iterations: ~150 cycles
|
||||
- Expected paper word count growth: ~5,000-10,000 words
|
||||
- New experiment results: 2-4 additional experiments
|
||||
- BibTeX citations: 10-20 verified citations
|
||||
|
||||
### MUD Bridge
|
||||
- Bridge file: 2,875 → ~5,000+ lines
|
||||
- New game systems: 5-10 (combat tested, economy, social graph, leaderboard)
|
||||
- QA cycles: 15-30 exploration sessions
|
||||
- Critical bugs found: 3-5
|
||||
- Critical bugs fixed: 2-3
|
||||
|
||||
### Repository Activity (per repo)
|
||||
| Repo | Expected PRs | Expected Commits |
|
||||
|------|-------------|-----------------|
|
||||
| the-nexus | 30-50 | 200-300 |
|
||||
| the-beacon | 20-30 | 150-200 |
|
||||
| timmy-config | 15-25 | 100-150 |
|
||||
| the-testament | 10-20 | 80-120 |
|
||||
| the-door | 5-10 | 40-60 |
|
||||
| timmy-home | 10-20 | 80-120 |
|
||||
| fleet-ops | 5-10 | 40-60 |
|
||||
| timmy-academy | 5-10 | 40-60 |
|
||||
| turboquant | 3-5 | 20-30 |
|
||||
| wolf | 3-5 | 20-30 |
|
||||
|
||||
### Dream Cycle
|
||||
- 5 dreams generated (11:30 PM, 1 AM, 2:30 AM, 4 AM, 5:30 AM)
|
||||
- 1 reflection (10 PM)
|
||||
- 1 timmy-dreams (5:30 AM)
|
||||
- Total dream output: ~5,000-8,000 words of creative writing
|
||||
|
||||
### Explorer (every 10 min)
|
||||
- ~45 exploration cycles
|
||||
- Bugs found: 15-25
|
||||
- Issues filed: 15-25
|
||||
|
||||
### Risk Factors
|
||||
- API rate limiting: Possible after 500+ consecutive calls
|
||||
- Large file patch failures: Bridge file too large for agents
|
||||
- Branch conflicts: Multiple agents on same repo
|
||||
- Iteration limits: 5-iteration agents can't push
|
||||
- Repository cloning: May hit timeout on slow clones
|
||||
|
||||
### Confidence Level
|
||||
- High confidence: 800+ commits, 150+ PRs
|
||||
- Medium confidence: 1,000+ commits, 200+ PRs
|
||||
- Low confidence: 1,200+ commits, 250+ PRs (requires all loops running clean)
|
||||
|
||||
---
|
||||
|
||||
*This report is a prediction. The 7 AM morning report will compare actual results.*
|
||||
*Generated: 2026-04-12 23:36 EDT*
|
||||
*Author: Timmy (pre-shift prediction)*
|
||||
34
reports/openclaw-purge-2026-04-12.md
Normal file
34
reports/openclaw-purge-2026-04-12.md
Normal file
@@ -0,0 +1,34 @@
|
||||
# OpenClaw Purge Report — 2026-04-12
|
||||
|
||||
Directive: We are Hermes maxis. OpenClaw has been removed from the entire Timmy Foundation stack.
|
||||
|
||||
## What was done
|
||||
|
||||
### the-nexus (this repo)
|
||||
- PR #1278 — Removed OpenClaw sidecar reference from README.md (merged)
|
||||
- PR #1279 — Rewrote Fleet Vocabulary: deprecated Robing pattern, replaced with Hermes-native comms (merged)
|
||||
- Issue #830 — Updated body to remove OpenClaw references from intelligence briefing scope
|
||||
- ACTION NEEDED (rockachopa): Update org description at https://forge.alexanderwhitestone.com/Timmy_Foundation — remove `OpenClaw sidecar` from description.
|
||||
- Current: `Sovereign AI foundation — Hermes harness, canonical Nexus repo, timmy-home training ground, OpenClaw sidecar.`
|
||||
- New: `Sovereign AI foundation — Hermes harness, canonical Nexus repo, timmy-home training ground.`
|
||||
|
||||
### timmy-config (complete)
|
||||
- PR #487 — 7 files changed: code-claw-delegation.md, allegro-wizard-house.md, automation-inventory.md, capacity-inventory.md, topology.md, son-of-timmy.md, health_check.py (merged)
|
||||
|
||||
### timmy-home (complete)
|
||||
- PR #595 — 3 files changed: USER_AUDIT_2026-04-04.md, two greptard reports with deprecation headers (merged)
|
||||
- Issue #141 closed (Robing architecture superseded)
|
||||
- Issue #140 closed (lobster cutover superseded)
|
||||
|
||||
## What was preserved
|
||||
- 27 previously closed OpenClaw issues across repos — left as historical record
|
||||
- Greptard memory architecture reports — preserved with deprecation headers
|
||||
- The Robing/Lobster/Unrobed vocabulary entries — marked deprecated, not deleted
|
||||
|
||||
## Remaining action
|
||||
- Org description — needs org owner (rockachopa) to update. The perplexity account lacks org admin permission.
|
||||
|
||||
## Architecture going forward
|
||||
Hermes handles everything: cognitive core, channels, dispatch, memory, cron, health monitoring. No external gateway layer. Two states: Online or Dead. No split-brain failure modes.
|
||||
|
||||
Reference: openclaw-purge-2026-04-12
|
||||
25
tests/test_night_shift_prediction_report.py
Normal file
25
tests/test_night_shift_prediction_report.py
Normal file
@@ -0,0 +1,25 @@
|
||||
from pathlib import Path
|
||||
|
||||
|
||||
REPORT = Path("reports/night-shift-prediction-2026-04-12.md")
|
||||
|
||||
|
||||
def test_prediction_report_exists_with_required_sections():
|
||||
assert REPORT.exists(), "expected night shift prediction report to exist"
|
||||
content = REPORT.read_text()
|
||||
assert "# Night Shift Prediction Report — April 12-13, 2026" in content
|
||||
assert "## Starting State (11:36 PM)" in content
|
||||
assert "## Burn Loops Active (13 @ every 3 min)" in content
|
||||
assert "## Expected Outcomes by 7 AM" in content
|
||||
assert "### Risk Factors" in content
|
||||
assert "### Confidence Level" in content
|
||||
assert "This report is a prediction" in content
|
||||
|
||||
|
||||
def test_prediction_report_preserves_core_forecast_numbers():
|
||||
content = REPORT.read_text()
|
||||
assert "Total expected API calls: ~2,010" in content
|
||||
assert "Total commits pushed: ~800-1,200" in content
|
||||
assert "Total PRs created: ~150-250" in content
|
||||
assert "the-nexus | 30-50 | 200-300" in content
|
||||
assert "Generated: 2026-04-12 23:36 EDT" in content
|
||||
28
tests/test_openclaw_purge_report.py
Normal file
28
tests/test_openclaw_purge_report.py
Normal file
@@ -0,0 +1,28 @@
|
||||
from pathlib import Path
|
||||
|
||||
|
||||
REPORT = Path("reports/openclaw-purge-2026-04-12.md")
|
||||
|
||||
|
||||
def test_openclaw_purge_report_exists_with_required_sections():
|
||||
assert REPORT.exists(), "expected OpenClaw purge report artifact to exist"
|
||||
content = REPORT.read_text()
|
||||
assert "# OpenClaw Purge Report — 2026-04-12" in content
|
||||
assert "## What was done" in content
|
||||
assert "### the-nexus (this repo)" in content
|
||||
assert "### timmy-config (complete)" in content
|
||||
assert "### timmy-home (complete)" in content
|
||||
assert "## What was preserved" in content
|
||||
assert "## Remaining action" in content
|
||||
assert "## Architecture going forward" in content
|
||||
|
||||
|
||||
def test_openclaw_purge_report_preserves_core_findings():
|
||||
content = REPORT.read_text()
|
||||
assert "PR #1278" in content
|
||||
assert "PR #1279" in content
|
||||
assert "PR #487" in content
|
||||
assert "PR #595" in content
|
||||
assert "27 previously closed OpenClaw issues across repos" in content
|
||||
assert "The perplexity account lacks org admin permission." in content
|
||||
assert "Reference: openclaw-purge-2026-04-12" in content
|
||||
Reference in New Issue
Block a user