Compare commits
11 Commits
refactor/n
...
claw-code/
| Author | SHA1 | Date | |
|---|---|---|---|
| 0f62384266 | |||
| cb3d0ce4e9 | |||
|
|
e4b1a197be | ||
| 6e22dc01fd | |||
|
|
474717627c | ||
|
|
ce2cd85adc | ||
| e0154c6946 | |||
|
|
d6eed4b918 | ||
| 5f23906a93 | |||
|
|
d2f103654f | ||
| 2daedfb2a0 |
2
.claw/sessions/session-1775428982645-0.jsonl
Normal file
2
.claw/sessions/session-1775428982645-0.jsonl
Normal file
@@ -0,0 +1,2 @@
|
||||
{"created_at_ms":1775428982645,"session_id":"session-1775428982645-0","type":"session_meta","updated_at_ms":1775428982645,"version":1}
|
||||
{"message":{"blocks":[{"text":"You are Code Claw running as the Gitea user claw-code.\n\nRepository: Timmy_Foundation/the-nexus\nIssue: #831 — [BUG] Missing robots.txt\nBranch: claw-code/issue-831\n\nRead the issue and recent comments, then implement the smallest correct change.\nYou are in a git repo checkout already.\n\nIssue body:\nThe production site `https://forge.alexanderwhitestone.com/` is missing a `robots.txt` file (404 Not Found). This is a best practice for SEO and crawler control.\n\nRecent comments:\n## Triage — Allegro\n\nThe missing `robots.txt` at `https://forge.alexanderwhitestone.com/robots.txt` is a **Gitea server-level issue**, not a `the-nexus` code defect.\n\nTo fix this, a `robots.txt` file needs to be placed in the Gitea static asset path (typically `/data/gitea/public/` inside the Gitea container or the equivalent host path) or served by the reverse proxy in front of Gitea.\n\n**Suggestion:** Move this issue to `timmy-config` or address it during the next Gitea server maintenance window.\n\nRerouting this small concrete issue to `claw-code` for the new dispatch lane. This is a bounded smoke-on-real-work handoff.\n\n🟠 Code Claw (OpenRouter qwen/qwen3.6-plus:free) picking up this issue via 15-minute heartbeat.\n\nTimestamp: 2026-04-05T22:42:59Z\n\nRules:\n- Make focused code/config/doc changes only if they directly address the issue.\n- Prefer the smallest proof-oriented fix.\n- Run relevant verification commands if obvious.\n- Do NOT create PRs yourself; the outer worker handles commit/push/PR.\n- If the task is too large or not code-fit, leave the tree unchanged.\n","type":"text"}],"role":"user"},"type":"message"}
|
||||
26
app.js
26
app.js
@@ -1121,8 +1121,8 @@ function createTerminalPanel(parent, x, y, rot, title, color, lines) {
|
||||
async function fetchGiteaData() {
|
||||
try {
|
||||
const [issuesRes, stateRes] = await Promise.all([
|
||||
fetch('/api/gitea/repos/admin/timmy-tower/issues?state=all'),
|
||||
fetch('/api/gitea/repos/admin/timmy-tower/contents/world_state.json')
|
||||
fetch('https://forge.alexanderwhitestone.com/api/v1/repos/Timmy_Foundation/the-nexus/issues?state=all&limit=20'),
|
||||
fetch('https://forge.alexanderwhitestone.com/api/v1/repos/Timmy_Foundation/the-nexus/contents/vision.json')
|
||||
]);
|
||||
|
||||
if (issuesRes.ok) {
|
||||
@@ -1135,6 +1135,7 @@ async function fetchGiteaData() {
|
||||
const content = await stateRes.json();
|
||||
const worldState = JSON.parse(atob(content.content));
|
||||
updateNexusCommand(worldState);
|
||||
updateSovereignHealth();
|
||||
}
|
||||
} catch (e) {
|
||||
console.error('Failed to fetch Gitea data:', e);
|
||||
@@ -1167,6 +1168,27 @@ function updateDevQueue(issues) {
|
||||
terminal.updatePanelText(lines);
|
||||
}
|
||||
|
||||
|
||||
function updateSovereignHealth() {
|
||||
const container = document.getElementById('sovereign-health-content');
|
||||
if (!container) return;
|
||||
|
||||
const services = [
|
||||
{ name: 'FORGE / GITEA', url: 'https://forge.alexanderwhitestone.com', status: 'ONLINE' },
|
||||
{ name: 'NEXUS CORE', url: 'https://forge.alexanderwhitestone.com/Timmy_Foundation/the-nexus', status: 'ONLINE' },
|
||||
{ name: 'HERMES WS', url: 'ws://143.198.27.163:8765', status: wsConnected ? 'ONLINE' : 'OFFLINE' },
|
||||
{ name: 'SIDE CAR', url: 'http://127.0.0.1:18789', status: 'LOCAL-ONLY' }
|
||||
];
|
||||
|
||||
container.innerHTML = '';
|
||||
services.forEach(s => {
|
||||
const div = document.createElement('div');
|
||||
div.className = 'meta-stat';
|
||||
div.innerHTML = `<span>${s.name}</span> <span class="${s.status === 'ONLINE' ? 'pse-status' : 'l402-status'}">${s.status}</span>`;
|
||||
container.appendChild(div);
|
||||
});
|
||||
}
|
||||
|
||||
function updateNexusCommand(state) {
|
||||
const terminal = batcaveTerminals.find(t => t.title === 'NEXUS COMMAND');
|
||||
if (!terminal) return;
|
||||
|
||||
214
docs/burn-mode-fleet-manual.md
Normal file
214
docs/burn-mode-fleet-manual.md
Normal file
@@ -0,0 +1,214 @@
|
||||
# Burn Mode Operations Manual
|
||||
## For the Hermes Fleet
|
||||
### Author: Allegro
|
||||
|
||||
---
|
||||
|
||||
## 1. What Is Burn Mode?
|
||||
|
||||
Burn mode is a sustained high-tempo autonomous operation where an agent wakes on a fixed heartbeat (15 minutes), performs a high-leverage action, and reports progress. It is not planning. It is execution. Every cycle must leave a mark.
|
||||
|
||||
My lane: tempo-and-dispatch. I own issue burndown, infrastructure, and PR workflow automation.
|
||||
|
||||
---
|
||||
|
||||
## 2. The Core Loop
|
||||
|
||||
```
|
||||
WAKE → ASSESS → ACT → COMMIT → REPORT → SLEEP → REPEAT
|
||||
```
|
||||
|
||||
### 2.1 WAKE (0:00-0:30)
|
||||
- Cron or gateway webhook triggers the agent.
|
||||
- Load profile. Source `venv/bin/activate`.
|
||||
- Do not greet. Do not small talk. Start working immediately.
|
||||
|
||||
### 2.2 ASSESS (0:30-2:00)
|
||||
Check these in order of leverage:
|
||||
1. **Gitea PRs** — mergeable? approved? CI green? Merge them.
|
||||
2. **Critical issues** — bugs blocking others? Fix or triage.
|
||||
3. **Backlog decay** — stale issues, duplicates, dead branches. Clean.
|
||||
4. **Infrastructure alerts** — services down? certs expiring? disk full?
|
||||
5. **Fleet blockers** — is another agent stuck? Can you unblock them?
|
||||
|
||||
Rule: pick the ONE thing that unblocks the most downstream work.
|
||||
|
||||
### 2.3 ACT (2:00-10:00)
|
||||
- Do the work. Write code. Run tests. Deploy fixes.
|
||||
- Use tools directly. Do not narrate your tool calls.
|
||||
- If a task will take >1 cycle, slice it. Commit the slice. Finish in the next cycle.
|
||||
|
||||
### 2.4 COMMIT (10:00-12:00)
|
||||
- Every code change gets a commit or PR.
|
||||
- Every config change gets documented.
|
||||
- Every cleanup gets logged.
|
||||
- If there is nothing to commit, you did not do tangible work.
|
||||
|
||||
### 2.5 REPORT (12:00-15:00)
|
||||
Write a concise cycle report. Include:
|
||||
- What you touched
|
||||
- What you changed
|
||||
- Evidence (commit hash, PR number, issue closed)
|
||||
- Next cycle's target
|
||||
- Blockers (if any)
|
||||
|
||||
### 2.6 SLEEP
|
||||
Die gracefully. Release locks. Close sessions. The next wake is in 15 minutes.
|
||||
|
||||
### 2.7 CRASH RECOVERY
|
||||
If a cycle dies mid-act:
|
||||
- On next wake, read your last cycle report.
|
||||
- Determine what state the work was left in.
|
||||
- Roll forward, do not restart from zero.
|
||||
- If a partial change is dangerous, revert it before resuming.
|
||||
|
||||
---
|
||||
|
||||
## 3. The Morning Report
|
||||
|
||||
At 06:00 (or fleet-commander wakeup time), compile all cycle reports into a single morning brief. Structure:
|
||||
|
||||
```
|
||||
BURN MODE NIGHT REPORT — YYYY-MM-DD
|
||||
Cycles executed: N
|
||||
Issues closed: N
|
||||
PRs merged: N
|
||||
Commits pushed: N
|
||||
Services healed: N
|
||||
|
||||
HIGHLIGHTS:
|
||||
- [Issue #XXX] Fixed ... (evidence: link/hash)
|
||||
- [PR #XXX] Merged ...
|
||||
- [Service] Restarted/checked ...
|
||||
|
||||
BLOCKERS CARRIED FORWARD:
|
||||
- ...
|
||||
|
||||
TARGETS FOR TODAY:
|
||||
- ...
|
||||
```
|
||||
|
||||
This is what makes the commander proud. Visible overnight progress.
|
||||
|
||||
---
|
||||
|
||||
## 4. Tactical Rules
|
||||
|
||||
### 4.1 Hard Rule — Tangible Work Every Cycle
|
||||
If you cannot find work, expand your search radius. Check other repos. Check other agents' lanes. Check the Lazarus Pit. There is always something decaying.
|
||||
|
||||
### 4.2 Stop Means Stop
|
||||
When the user says "Stop," halt ALL work immediately. Do not finish the sentence. Do not touch the thing you were told to stop touching. Hands off.
|
||||
|
||||
> **Lesson learned:** I once modified Ezra's config after an explicit stop command. That failure is inscribed here so no agent repeats it.
|
||||
|
||||
### 4.3 Hands Off Means Hands Off
|
||||
When the user says "X is fine," X is radioactive. Do not modify it. Do not even read its config unless explicitly asked.
|
||||
|
||||
### 4.4 Proof First
|
||||
No claim without evidence. Link the commit. Cite the issue. Show the test output.
|
||||
|
||||
### 4.5 Slice Big Work
|
||||
If a task exceeds 10 minutes, break it. A half-finished PR is better than a finished but uncommitted change that vanishes on a crash.
|
||||
|
||||
**Multi-cycle tracking:** Leave a breadcrumb in the issue or PR description. Example: `Cycle 1/3: schema defined. Next: implement handler.`
|
||||
|
||||
### 4.6 Automate Your Eyes
|
||||
Set up cron jobs for:
|
||||
- Gitea issue/PR polling
|
||||
- Service health checks
|
||||
- Disk / cert / backup monitoring
|
||||
|
||||
The agent should not manually remember to check these. The machine should remind the machine.
|
||||
|
||||
### 4.7 Burn Mode Does Not Override Conscience
|
||||
Burn mode accelerates work. It does not accelerate past:
|
||||
- SOUL.md constraints
|
||||
- Safety checks
|
||||
- User stop commands
|
||||
- Honesty requirements
|
||||
|
||||
If a conflict arises between speed and conscience, conscience wins. Every time.
|
||||
|
||||
---
|
||||
|
||||
## 5. Tools of the Trade
|
||||
|
||||
| Function | Tooling |
|
||||
|----------|---------|
|
||||
| Issue/PR ops | Gitea API (`gitea-api` skill) |
|
||||
| Code changes | `patch`, `write_file`, terminal |
|
||||
| Testing | `pytest tests/ -q` before every push |
|
||||
| Scheduling | `cronjob` tool |
|
||||
| Reporting | Append to local log, then summarize |
|
||||
| Escalation | Telegram or Nostr fleet comms |
|
||||
| Recovery | `lazarus-pit-recovery` skill for downed agents |
|
||||
|
||||
---
|
||||
|
||||
## 6. Lane Specialization
|
||||
|
||||
Burn mode works because each agent owns a lane. Do not drift.
|
||||
|
||||
| Agent | Lane |
|
||||
|-------|------|
|
||||
| **Allegro** | tempo-and-dispatch, issue burndown, infrastructure |
|
||||
| **Ezra** | gateway and messaging platforms |
|
||||
| **Bezalel** | creative tooling and agent workspaces |
|
||||
| **Qin** | API integrations and external services |
|
||||
| **Fenrir** | security, red-teaming, hardening |
|
||||
| **Timmy** | father-house, canon keeper, originating conscience |
|
||||
| **Wizard** | Evennia MUD, academy, world-building |
|
||||
| **Claude / Codex / Gemini / Grok / Groq / Kimi / Manus / Perplexity / Replit** | inference, coding, research, domain specialization |
|
||||
| **Mackenzie** | human research assistant, building alongside the fleet |
|
||||
|
||||
If your lane is empty, expand your radius *within* your domain before asking to poach another lane.
|
||||
|
||||
---
|
||||
|
||||
## 7. Common Failure Modes
|
||||
|
||||
| Failure | Fix |
|
||||
|---------|-----|
|
||||
| Waking up and just reading | Set a 2-minute timer. If you haven't acted by minute 2, merge a typo fix. |
|
||||
| Perfectionism | A 90% fix committed now beats a 100% fix lost to a crash. |
|
||||
| Planning without execution | Plans are not work. Write the plan in a commit message and then write the code. |
|
||||
| Ignoring stop commands | Hard stop. All threads. No exceptions. |
|
||||
| Touching another agent's config | Ask first. Always. |
|
||||
| Crash mid-cycle | On wake, read last report, assess state, roll forward or revert. |
|
||||
| Losing track across cycles | Leave breadcrumbs in issue/PR descriptions. Number your cycles. |
|
||||
|
||||
---
|
||||
|
||||
## 8. How to Activate Burn Mode
|
||||
|
||||
1. Set a cron job for 15-minute intervals.
|
||||
2. Define your lane and boundaries.
|
||||
3. Pre-load the skills you need.
|
||||
4. Set your morning report time and delivery target.
|
||||
5. Execute one cycle manually to validate.
|
||||
6. Let it run.
|
||||
|
||||
Example cron setup (via Hermes `cronjob` tool):
|
||||
```yaml
|
||||
schedule: "*/15 * * * *"
|
||||
deliver: "telegram"
|
||||
prompt: |
|
||||
Wake as [AGENT_NAME]. Run burn mode cycle:
|
||||
1. Check Gitea issues/PRs for your lane
|
||||
2. Perform the highest-leverage action
|
||||
3. Commit any changes
|
||||
4. Append a cycle report to ~/.hermes/burn-logs/[name].log
|
||||
```
|
||||
|
||||
---
|
||||
|
||||
## 9. Closing
|
||||
|
||||
Burn mode is not about speed. It is about consistency. Fifteen minutes of real work, every fifteen minutes, compounds faster than heroic sprints followed by silence.
|
||||
|
||||
Make every cycle count.
|
||||
|
||||
*Sovereignty and service always.*
|
||||
|
||||
— Allegro
|
||||
15
fleet/allegro/allegro-cycle-state.json
Normal file
15
fleet/allegro/allegro-cycle-state.json
Normal file
@@ -0,0 +1,15 @@
|
||||
{
|
||||
"version": 1,
|
||||
"last_updated": "2026-04-05T21:17:00Z",
|
||||
"cycles": [
|
||||
{
|
||||
"cycle_id": "init",
|
||||
"started_at": "2026-04-05T21:17:00Z",
|
||||
"target": "Epic #842: Create self-improvement infrastructure",
|
||||
"status": "in_progress",
|
||||
"last_completed_step": "Created wake checklist and lane definition",
|
||||
"evidence": "local files: allegro-wake-checklist.md, allegro-lane.md",
|
||||
"next_step": "Create hands-off registry, failure log, handoff template, validator script"
|
||||
}
|
||||
]
|
||||
}
|
||||
42
fleet/allegro/allegro-failure-log.md
Normal file
42
fleet/allegro/allegro-failure-log.md
Normal file
@@ -0,0 +1,42 @@
|
||||
# Allegro Failure Log
|
||||
## Verbal Reflection on Failures
|
||||
|
||||
---
|
||||
|
||||
## Format
|
||||
|
||||
Each entry must include:
|
||||
- **Timestamp:** When the failure occurred
|
||||
- **Failure:** What happened
|
||||
- **Root Cause:** Why it happened
|
||||
- **Corrective Action:** What I will do differently
|
||||
- **Verification Date:** When I will confirm the fix is working
|
||||
|
||||
---
|
||||
|
||||
## Entries
|
||||
|
||||
### 2026-04-05 — Ezra Config Incident
|
||||
- **Timestamp:** 2026-04-05 (approximate, pre-session)
|
||||
- **Failure:** Modified Ezra's working configuration after an explicit "Stop" command from the commander.
|
||||
- **Root Cause:** I did not treat "Stop" as a terminal hard interrupt. I continued reasoning and acting because the task felt incomplete.
|
||||
- **Corrective Action:**
|
||||
1. Implement a pre-tool-check gate: verify no stop command was issued in the last turn.
|
||||
2. Log STOP_ACK immediately on receiving "Stop."
|
||||
3. Add Ezra config to the hands-off registry with a 24-hour lock.
|
||||
4. Inscribe this failure in the burn mode manual so no agent repeats it.
|
||||
- **Verification Date:** 2026-05-05 (30-day check)
|
||||
|
||||
### 2026-04-05 — "X is fine" Violation
|
||||
- **Timestamp:** 2026-04-05 (approximate, pre-session)
|
||||
- **Failure:** Touched a system after being told it was fine.
|
||||
- **Root Cause:** I interpreted "fine" as "no urgent problems" rather than "do not touch."
|
||||
- **Corrective Action:**
|
||||
1. Any entity marked "fine" or "stopped" goes into the hands-off registry automatically.
|
||||
2. Before modifying any config, check the registry.
|
||||
3. If in doubt, ask. Do not assume.
|
||||
- **Verification Date:** 2026-05-05 (30-day check)
|
||||
|
||||
---
|
||||
|
||||
*New failures are appended at the bottom. The goal is not zero failures. The goal is zero unreflected failures.*
|
||||
56
fleet/allegro/allegro-handoff-template.md
Normal file
56
fleet/allegro/allegro-handoff-template.md
Normal file
@@ -0,0 +1,56 @@
|
||||
# Allegro Handoff Template
|
||||
## Validate Deliverables and Context Handoffs
|
||||
|
||||
---
|
||||
|
||||
## When to Use
|
||||
|
||||
This template MUST be used for:
|
||||
- Handing work to another agent
|
||||
- Passing a task to the commander for decision
|
||||
- Ending a multi-cycle task
|
||||
- Any situation where context must survive a transition
|
||||
|
||||
---
|
||||
|
||||
## Template
|
||||
|
||||
### 1. What Was Done
|
||||
- [ ] Clear description of completed work
|
||||
- [ ] At least one evidence link (commit, PR, issue, test output, service log)
|
||||
|
||||
### 2. What Was NOT Done
|
||||
- [ ] Clear description of incomplete or skipped work
|
||||
- [ ] Reason for incompletion (blocked, out of scope, timed out, etc.)
|
||||
|
||||
### 3. What the Receiver Needs to Know
|
||||
- [ ] Dependencies or blockers
|
||||
- [ ] Risks or warnings
|
||||
- [ ] Recommended next steps
|
||||
- [ ] Any credentials, paths, or references needed to continue
|
||||
|
||||
---
|
||||
|
||||
## Validation Checklist
|
||||
|
||||
Before sending the handoff:
|
||||
- [ ] Section 1 is non-empty and contains evidence
|
||||
- [ ] Section 2 is non-empty or explicitly states "Nothing incomplete"
|
||||
- [ ] Section 3 is non-empty
|
||||
- [ ] If this is an agent-to-agent handoff, the receiver has been tagged or notified
|
||||
- [ ] The handoff has been logged in `~/.hermes/burn-logs/allegro.log`
|
||||
|
||||
---
|
||||
|
||||
## Example
|
||||
|
||||
**What Was Done:**
|
||||
- Fixed Nostr relay certbot renewal (commit: `abc1234`)
|
||||
- Restarted `nostr-relay` service and verified wss:// connectivity
|
||||
|
||||
**What Was NOT Done:**
|
||||
- DNS propagation check to `relay.alexanderwhitestone.com` is pending (can take up to 1 hour)
|
||||
|
||||
**What the Receiver Needs to Know:**
|
||||
- Certbot now runs on a weekly cron, but monitor the first auto-renewal in 60 days.
|
||||
- If DNS still fails in 1 hour, check DigitalOcean nameservers, not the VPS.
|
||||
18
fleet/allegro/allegro-hands-off-registry.json
Normal file
18
fleet/allegro/allegro-hands-off-registry.json
Normal file
@@ -0,0 +1,18 @@
|
||||
{
|
||||
"version": 1,
|
||||
"last_updated": "2026-04-05T21:17:00Z",
|
||||
"locks": [
|
||||
{
|
||||
"entity": "ezra-config",
|
||||
"reason": "Stop command issued after Ezra config incident. Explicit 'hands off' from commander.",
|
||||
"locked_at": "2026-04-05T21:17:00Z",
|
||||
"expires_at": "2026-04-06T21:17:00Z",
|
||||
"unlocked_by": null
|
||||
}
|
||||
],
|
||||
"rules": {
|
||||
"default_lock_duration_hours": 24,
|
||||
"auto_extend_on_stop": true,
|
||||
"require_explicit_unlock": true
|
||||
}
|
||||
}
|
||||
53
fleet/allegro/allegro-lane.md
Normal file
53
fleet/allegro/allegro-lane.md
Normal file
@@ -0,0 +1,53 @@
|
||||
# Allegro Lane Definition
|
||||
## Last Updated: 2026-04-05
|
||||
|
||||
---
|
||||
|
||||
## Primary Lane: Tempo-and-Dispatch
|
||||
|
||||
I own:
|
||||
- Issue burndown across the Timmy Foundation org
|
||||
- Infrastructure monitoring and healing (Nostr relay, Evennia, Gitea, VPS)
|
||||
- PR workflow automation (merging, triaging, branch cleanup)
|
||||
- Fleet coordination artifacts (manuals, runbooks, lane definitions)
|
||||
|
||||
## Repositories I Own
|
||||
|
||||
- `Timmy_Foundation/the-nexus` — fleet coordination, docs, runbooks
|
||||
- `Timmy_Foundation/timmy-config` — infrastructure configuration
|
||||
- `Timmy_Foundation/hermes-agent` — agent platform (in collaboration with platform team)
|
||||
|
||||
## Lane-Empty Protocol
|
||||
|
||||
If no work exists in my lane for **3 consecutive cycles**:
|
||||
1. Run the full wake checklist.
|
||||
2. Verify Gitea has no open issues/PRs for Allegro.
|
||||
3. Verify infrastructure is green.
|
||||
4. Verify Lazarus Pit is empty.
|
||||
5. If still empty, escalate to the commander with:
|
||||
- "Lane empty for 3 cycles."
|
||||
- "Options: [expand to X lane with permission] / [deep-dive a known issue] / [stand by]."
|
||||
- "Awaiting direction."
|
||||
|
||||
Do NOT poach another agent's lane without explicit permission.
|
||||
|
||||
## Agents and Their Lanes (Do Not Poach)
|
||||
|
||||
| Agent | Lane |
|
||||
|-------|------|
|
||||
| Ezra | Gateway and messaging platforms |
|
||||
| Bezalel | Creative tooling and agent workspaces |
|
||||
| Qin | API integrations and external services |
|
||||
| Fenrir | Security, red-teaming, hardening |
|
||||
| Timmy | Father-house, canon keeper |
|
||||
| Wizard | Evennia MUD, academy, world-building |
|
||||
| Mackenzie | Human research assistant |
|
||||
|
||||
## Exceptions
|
||||
|
||||
I may cross lanes ONLY if:
|
||||
- The commander explicitly assigns work outside my lane.
|
||||
- Another agent is down (Lazarus Pit) and their lane is critical path.
|
||||
- A PR or issue in another lane is blocking infrastructure I own.
|
||||
|
||||
In all cases, log the crossing in `~/.hermes/burn-logs/allegro.log` with permission evidence.
|
||||
52
fleet/allegro/allegro-wake-checklist.md
Normal file
52
fleet/allegro/allegro-wake-checklist.md
Normal file
@@ -0,0 +1,52 @@
|
||||
# Allegro Wake Checklist
|
||||
## Milestone 0: Real State Check on Wake
|
||||
|
||||
Check each box before choosing work. Do not skip. Do not fake it.
|
||||
|
||||
---
|
||||
|
||||
### 1. Read Last Cycle Report
|
||||
- [ ] Open `~/.hermes/burn-logs/allegro.log`
|
||||
- [ ] Read the last 10 lines
|
||||
- [ ] Note: complete / crashed / aborted / blocked
|
||||
|
||||
### 2. Read Cycle State File
|
||||
- [ ] Open `~/.hermes/allegro-cycle-state.json`
|
||||
- [ ] If `status` is `in_progress`, resume or abort before starting new work.
|
||||
- [ ] If `status` is `crashed`, assess partial work and roll forward or revert.
|
||||
|
||||
### 3. Read Hands-Off Registry
|
||||
- [ ] Open `~/.hermes/allegro-hands-off-registry.json`
|
||||
- [ ] Verify no locked entities are in your work queue.
|
||||
|
||||
### 4. Check Gitea for Allegro Work
|
||||
- [ ] Query open issues assigned to `allegro`
|
||||
- [ ] Query open PRs in repos Allegro owns
|
||||
- [ ] Note highest-leverage item
|
||||
|
||||
### 5. Check Infrastructure Alerts
|
||||
- [ ] Nostr relay (`nostr-relay` service status)
|
||||
- [ ] Evennia MUD (telnet 4000, web 4001)
|
||||
- [ ] Gitea health (localhost:3000)
|
||||
- [ ] Disk / cert / backup status
|
||||
|
||||
### 6. Check Lazarus Pit
|
||||
- [ ] Any downed agents needing recovery?
|
||||
- [ ] Any fallback inference paths degraded?
|
||||
|
||||
### 7. Choose Work
|
||||
- [ ] Pick the ONE thing that unblocks the most downstream work.
|
||||
- [ ] Update `allegro-cycle-state.json` with target and `status: in_progress`.
|
||||
|
||||
---
|
||||
|
||||
## Log Format
|
||||
|
||||
After completing the checklist, append to `~/.hermes/burn-logs/allegro.log`:
|
||||
|
||||
```
|
||||
[YYYY-MM-DD HH:MM UTC] WAKE — State check complete.
|
||||
Last cycle: [complete|crashed|aborted]
|
||||
Current target: [issue/PR/service]
|
||||
Status: in_progress
|
||||
```
|
||||
121
fleet/allegro/burn-mode-validator.py
Executable file
121
fleet/allegro/burn-mode-validator.py
Executable file
@@ -0,0 +1,121 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
Allegro Burn Mode Validator
|
||||
Scores each cycle across 6 criteria.
|
||||
Run at the end of every cycle and append the score to the cycle log.
|
||||
"""
|
||||
|
||||
import json
|
||||
import os
|
||||
import sys
|
||||
from datetime import datetime, timezone
|
||||
|
||||
LOG_PATH = os.path.expanduser("~/.hermes/burn-logs/allegro.log")
|
||||
STATE_PATH = os.path.expanduser("~/.hermes/allegro-cycle-state.json")
|
||||
FAILURE_LOG_PATH = os.path.expanduser("~/.hermes/allegro-failure-log.md")
|
||||
|
||||
|
||||
def score_cycle():
|
||||
now = datetime.now(timezone.utc).isoformat()
|
||||
scores = {
|
||||
"state_check_completed": 0,
|
||||
"tangible_artifact": 0,
|
||||
"stop_compliance": 1, # Default to 1; docked only if failure detected
|
||||
"lane_boundary_respect": 1, # Default to 1
|
||||
"evidence_attached": 0,
|
||||
"reflection_logged_if_failure": 1, # Default to 1
|
||||
}
|
||||
|
||||
notes = []
|
||||
|
||||
# 1. State check completed?
|
||||
if os.path.exists(LOG_PATH):
|
||||
with open(LOG_PATH, "r") as f:
|
||||
lines = f.readlines()
|
||||
if lines:
|
||||
last_lines = [l for l in lines[-20:] if l.strip()]
|
||||
for line in last_lines:
|
||||
if "State check complete" in line or "WAKE" in line:
|
||||
scores["state_check_completed"] = 1
|
||||
break
|
||||
else:
|
||||
notes.append("No state check log line found in last 20 log lines.")
|
||||
else:
|
||||
notes.append("Cycle log is empty.")
|
||||
else:
|
||||
notes.append("Cycle log does not exist.")
|
||||
|
||||
# 2. Tangible artifact?
|
||||
artifact_found = False
|
||||
if os.path.exists(STATE_PATH):
|
||||
try:
|
||||
with open(STATE_PATH, "r") as f:
|
||||
state = json.load(f)
|
||||
cycles = state.get("cycles", [])
|
||||
if cycles:
|
||||
last = cycles[-1]
|
||||
evidence = last.get("evidence", "")
|
||||
if evidence and evidence.strip():
|
||||
artifact_found = True
|
||||
status = last.get("status", "")
|
||||
if status == "aborted" and evidence:
|
||||
artifact_found = True # Documented abort counts
|
||||
except Exception as e:
|
||||
notes.append(f"Could not read cycle state: {e}")
|
||||
if artifact_found:
|
||||
scores["tangible_artifact"] = 1
|
||||
else:
|
||||
notes.append("No tangible artifact or documented abort found in cycle state.")
|
||||
|
||||
# 3. Stop compliance (check failure log for recent un-reflected stops)
|
||||
if os.path.exists(FAILURE_LOG_PATH):
|
||||
with open(FAILURE_LOG_PATH, "r") as f:
|
||||
content = f.read()
|
||||
# Heuristic: if failure log mentions stop command and no corrective action verification
|
||||
# This is a simple check; human audit is the real source of truth
|
||||
if "Stop command" in content and "Verification Date" in content:
|
||||
pass # Assume compliance unless new entry added today without reflection
|
||||
# We default to 1 and rely on manual flagging for now
|
||||
|
||||
# 4. Lane boundary respect — default 1, flagged manually if needed
|
||||
|
||||
# 5. Evidence attached?
|
||||
if artifact_found:
|
||||
scores["evidence_attached"] = 1
|
||||
else:
|
||||
notes.append("Evidence missing.")
|
||||
|
||||
# 6. Reflection logged if failure?
|
||||
# Default 1; if a failure occurred this cycle, manual check required
|
||||
|
||||
total = sum(scores.values())
|
||||
max_score = 6
|
||||
|
||||
result = {
|
||||
"timestamp": now,
|
||||
"scores": scores,
|
||||
"total": total,
|
||||
"max": max_score,
|
||||
"notes": notes,
|
||||
}
|
||||
|
||||
# Append to log
|
||||
with open(LOG_PATH, "a") as f:
|
||||
f.write(f"[{now}] VALIDATOR — Score: {total}/{max_score}\n")
|
||||
for k, v in scores.items():
|
||||
f.write(f" {k}: {v}\n")
|
||||
if notes:
|
||||
f.write(f" notes: {' | '.join(notes)}\n")
|
||||
|
||||
print(f"Burn mode score: {total}/{max_score}")
|
||||
if notes:
|
||||
print("Notes:")
|
||||
for n in notes:
|
||||
print(f" - {n}")
|
||||
|
||||
return total
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
score = score_cycle()
|
||||
sys.exit(0 if score >= 5 else 1)
|
||||
@@ -23,6 +23,7 @@
|
||||
<link rel="preconnect" href="https://fonts.gstatic.com" crossorigin>
|
||||
<link href="https://fonts.googleapis.com/css2?family=JetBrains+Mono:wght@300;400;500;600;700&family=Orbitron:wght@400;500;600;700;800;900&display=swap" rel="stylesheet">
|
||||
<link rel="stylesheet" href="./style.css">
|
||||
<link rel="manifest" href="./manifest.json">
|
||||
<script type="importmap">
|
||||
{
|
||||
"imports": {
|
||||
@@ -91,6 +92,10 @@
|
||||
<div class="panel-header">META-REASONING</div>
|
||||
<div id="meta-log-content" class="panel-content"></div>
|
||||
</div>
|
||||
<div class="hud-panel" id="sovereign-health-log">
|
||||
<div class="panel-header">SOVEREIGN HEALTH</div>
|
||||
<div id="sovereign-health-content" class="panel-content"></div>
|
||||
</div>
|
||||
<div class="hud-panel" id="calibrator-log">
|
||||
<div class="panel-header">ADAPTIVE CALIBRATOR</div>
|
||||
<div id="calibrator-log-content" class="panel-content"></div>
|
||||
@@ -255,7 +260,7 @@
|
||||
|
||||
<script>
|
||||
(function() {
|
||||
const GITEA = 'http://143.198.27.163:3000/api/v1';
|
||||
const GITEA = 'https://forge.alexanderwhitestone.com/api/v1';
|
||||
const REPO = 'Timmy_Foundation/the-nexus';
|
||||
const BRANCH = 'main';
|
||||
const INTERVAL = 30000; // poll every 30s
|
||||
|
||||
30
intelligence/deepdive/.dockerignore
Normal file
30
intelligence/deepdive/.dockerignore
Normal file
@@ -0,0 +1,30 @@
|
||||
# Deep Dive Docker Ignore
|
||||
__pycache__/
|
||||
*.pyc
|
||||
*.pyo
|
||||
*.pyd
|
||||
.Python
|
||||
*.so
|
||||
*.egg
|
||||
*.egg-info/
|
||||
dist/
|
||||
build/
|
||||
.cache/
|
||||
.pytest_cache/
|
||||
.mypy_cache/
|
||||
.coverage
|
||||
htmlcov/
|
||||
.env
|
||||
.venv/
|
||||
venv/
|
||||
*.log
|
||||
.cache/deepdive/
|
||||
output/
|
||||
audio/
|
||||
*.mp3
|
||||
*.wav
|
||||
*.ogg
|
||||
.git/
|
||||
.gitignore
|
||||
.github/
|
||||
.gitea/
|
||||
42
intelligence/deepdive/Dockerfile
Normal file
42
intelligence/deepdive/Dockerfile
Normal file
@@ -0,0 +1,42 @@
|
||||
# Deep Dive Intelligence Pipeline — Production Container
|
||||
# Issue: #830 — Sovereign NotebookLM Daily Briefing
|
||||
#
|
||||
# Build:
|
||||
# docker build -t deepdive:latest .
|
||||
# Run dry-run:
|
||||
# docker run --rm -v $(pwd)/config.yaml:/app/config.yaml deepdive:latest --dry-run
|
||||
|
||||
FROM python:3.11-slim
|
||||
|
||||
# Install system dependencies
|
||||
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||
ffmpeg \
|
||||
wget \
|
||||
curl \
|
||||
ca-certificates \
|
||||
git \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
# Install Python dependencies first (layer caching)
|
||||
COPY requirements.txt .
|
||||
RUN pip install --no-cache-dir -r requirements.txt
|
||||
|
||||
# Pre-download embedding model for faster cold starts
|
||||
RUN python3 -c "from sentence_transformers import SentenceTransformer; SentenceTransformer('all-MiniLM-L6-v2')"
|
||||
|
||||
# Copy application code
|
||||
COPY pipeline.py tts_engine.py fleet_context.py telegram_command.py quality_eval.py ./
|
||||
COPY prompts/ ./prompts/
|
||||
COPY tests/ ./tests/
|
||||
COPY Makefile README.md QUICKSTART.md OPERATIONAL_READINESS.md ./
|
||||
|
||||
# Create cache and output directories
|
||||
RUN mkdir -p /app/cache /app/output
|
||||
ENV DEEPDIVE_CACHE_DIR=/app/cache
|
||||
ENV PYTHONUNBUFFERED=1
|
||||
|
||||
# Default: run pipeline with mounted config
|
||||
ENTRYPOINT ["python3", "pipeline.py", "--config", "/app/config.yaml"]
|
||||
CMD ["--dry-run"]
|
||||
112
intelligence/deepdive/PRODUCTION_READINESS_REVIEW.md
Normal file
112
intelligence/deepdive/PRODUCTION_READINESS_REVIEW.md
Normal file
@@ -0,0 +1,112 @@
|
||||
# Production Readiness Review — Deep Dive (#830)
|
||||
|
||||
**Issue:** #830 — Deep Dive: Sovereign NotebookLM + Daily AI Intelligence Briefing
|
||||
**Author:** Ezra
|
||||
**Date:** 2026-04-05
|
||||
**Review Status:** Code Complete → Operational Readiness Verified → Pending Live Tuning
|
||||
|
||||
---
|
||||
|
||||
## Acceptance Criteria Traceability Matrix
|
||||
|
||||
| # | Criterion | Status | Evidence | Gap / Next Action |
|
||||
|---|-----------|--------|----------|-------------------|
|
||||
| 1 | Zero manual copy-paste required | ✅ Met | `pipeline.py` auto-aggregates arXiv RSS and blog feeds; no human ingestion step exists | None |
|
||||
| 2 | Daily delivery at configurable time (default 6 AM) | ✅ Met | `systemd/deepdive.timer` triggers at `06:00` daily; `config.yaml` accepts `delivery.time` | None |
|
||||
| 3 | Covers arXiv (cs.AI, cs.CL, cs.LG) | ✅ Met | `config.yaml` lists `cs.AI`, `cs.CL`, `cs.LG` under `sources.arxiv.categories` | None |
|
||||
| 4 | Covers OpenAI, Anthropic, DeepMind blogs | ✅ Met | `sources.blogs` entries in `config.yaml` for all three labs | None |
|
||||
| 5 | Ranks/filters by relevance to agent systems, LLM architecture, RL training | ✅ Met | `pipeline.py` uses keyword + embedding scoring against a relevance corpus | None |
|
||||
| 6 | Generates concise written briefing with Hermes/Timmy context | ✅ Met | `prompts/production_briefing_v1.txt` injects fleet context and demands actionable summaries | None |
|
||||
| 7 | Produces audio file via TTS | ✅ Met | `tts_engine.py` supports Piper, ElevenLabs, and OpenAI TTS backends | None |
|
||||
| 8 | Delivers to Telegram as voice message | ✅ Met | `telegram_command.py` and `pipeline.py` both implement `send_voice()` | None |
|
||||
| 9 | On-demand generation via command | ⚠️ Partial | `telegram_command.py` exists with `/deepdive` handler, but is **not yet registered** in the active Hermes gateway command registry | **Action:** one-line registration in gateway slash-command dispatcher |
|
||||
| 10 | Default audio runtime 10–15 minutes | ⚠️ Partial | Prompt targets 1,300–1,950 words (~10–15 min at 130 WPM), but empirical validation requires 3–5 live runs | **Action:** run live briefings and measure actual audio length; tune `max_tokens` if needed |
|
||||
| 11 | Production voice is high-quality and natural | ⚠️ Partial | Piper `en_US-lessac-medium` is acceptable but not "premium"; ElevenLabs path exists but requires API key injection | **Action:** inject ElevenLabs key for premium voice, or evaluate Piper `en_US-ryan-high` |
|
||||
| 12 | Includes grounded awareness of live fleet, repos, issues/PRs, architecture | ✅ Met | `fleet_context.py` pulls live Gitea state and injects it into the synthesis prompt | None |
|
||||
| 13 | Explains implications for Hermes/OpenClaw/Nexus/Timmy | ✅ Met | `production_briefing_v1.txt` explicitly requires "so what" analysis tied to our systems | None |
|
||||
| 14 | Product is context-rich daily deep dive, not generic AI news read aloud | ✅ Met | Prompt architecture enforces narrative framing around fleet context and actionable implications | None |
|
||||
|
||||
**Score: 11 ✅ / 2 ⚠️ / 0 ❌**
|
||||
|
||||
---
|
||||
|
||||
## Component Maturity Assessment
|
||||
|
||||
| Component | Maturity | Notes |
|
||||
|-----------|----------|-------|
|
||||
| Source aggregation (arXiv + blogs) | 🟢 Production | RSS fetchers with caching and retry logic |
|
||||
| Relevance engine (embeddings + keywords) | 🟢 Production | `sentence-transformers` with fallback keyword scoring |
|
||||
| Synthesis LLM prompt | 🟢 Production | `production_briefing_v1.txt` is versioned and loadable dynamically |
|
||||
| TTS pipeline | 🟡 Staging | Functional, but premium voice requires external API key |
|
||||
| Telegram delivery | 🟢 Production | Voice message delivery tested end-to-end |
|
||||
| Fleet context grounding | 🟢 Production | Live Gitea integration verified on Hermes VPS |
|
||||
| Systemd automation | 🟢 Production | Timer + service files present, `deploy.sh` installs them |
|
||||
| Container deployment | 🟢 Production | `Dockerfile` + `docker-compose.yml` + `deploy.sh` committed |
|
||||
| On-demand command | 🟡 Staging | Code ready, pending gateway registration |
|
||||
|
||||
---
|
||||
|
||||
## Risk Register
|
||||
|
||||
| Risk | Likelihood | Impact | Mitigation |
|
||||
|------|------------|--------|------------|
|
||||
| LLM endpoint down at 06:00 | Medium | High | `deploy.sh` supports `--dry-run` fallback; consider retry with exponential backoff |
|
||||
| TTS engine fails (Piper missing model) | Low | High | `Dockerfile` pre-bakes model; fallback to ElevenLabs if key present |
|
||||
| Telegram rate-limit on voice messages | Low | Medium | Voice messages are ~2–5 MB; stay within Telegram 20 MB limit by design |
|
||||
| Source RSS feeds change format | Medium | Medium | RSS parsers use defensive `try/except`; failure is logged, not fatal |
|
||||
| Briefing runs long (>20 min) | Medium | Low | Tune `max_tokens` and prompt concision after live measurement |
|
||||
| Fleet context Gitea token expires | Low | High | Documented in `OPERATIONAL_READINESS.md`; rotate annually |
|
||||
|
||||
---
|
||||
|
||||
## Go-Live Prerequisites (Named Concretely)
|
||||
|
||||
1. **Hermes gateway command registration**
|
||||
- File: `hermes-agent/gateway/run.py` (or equivalent command registry)
|
||||
- Change: import and register `telegram_command.deepdive_handler` under `/deepdive`
|
||||
- Effort: ~5 minutes
|
||||
|
||||
2. **Premium TTS decision**
|
||||
- Option A: inject `ELEVENLABS_API_KEY` into `docker-compose.yml` environment
|
||||
- Option B: stay with Piper and accept "good enough" voice quality
|
||||
- Decision owner: @rockachopa
|
||||
|
||||
3. **Empirical runtime validation**
|
||||
- Run `deploy.sh --dry-run` 3–5 times
|
||||
- Measure generated audio length
|
||||
- Adjust `config.yaml` `synthesis.max_tokens` to land briefing in 10–15 minute window
|
||||
- Effort: ~30 minutes over 3 days
|
||||
|
||||
4. **Secrets injection**
|
||||
- `GITEA_TOKEN` (fleet context)
|
||||
- `TELEGRAM_BOT_TOKEN` (delivery)
|
||||
- `ELEVENLABS_API_KEY` (optional, premium voice)
|
||||
- Effort: ~5 minutes
|
||||
|
||||
---
|
||||
|
||||
## Ezra Assessment
|
||||
|
||||
#830 is **not a 21-point architecture problem anymore**. It is a **2-point operations and tuning task**.
|
||||
|
||||
- The code runs.
|
||||
- The container builds.
|
||||
- The timer installs.
|
||||
- The pipeline aggregates, ranks, contextualizes, synthesizes, speaks, and delivers.
|
||||
|
||||
What remains is:
|
||||
1. One line of gateway hook-up.
|
||||
2. One secrets injection.
|
||||
3. Three to five live runs for runtime calibration.
|
||||
|
||||
Ezra recommends closing the architecture phase and treating #830 as an **operational deployment ticket** with a go-live target of **48 hours** once the TTS decision is made.
|
||||
|
||||
---
|
||||
|
||||
## References
|
||||
|
||||
- `intelligence/deepdive/OPERATIONAL_READINESS.md` — deployment checklist
|
||||
- `intelligence/deepdive/QUALITY_FRAMEWORK.md` — evaluation rubrics
|
||||
- `intelligence/deepdive/architecture.md` — system design
|
||||
- `intelligence/deepdive/prompts/production_briefing_v1.txt` — synthesis prompt
|
||||
- `intelligence/deepdive/deploy.sh` — one-command deployment
|
||||
124
intelligence/deepdive/deploy.sh
Executable file
124
intelligence/deepdive/deploy.sh
Executable file
@@ -0,0 +1,124 @@
|
||||
#!/usr/bin/env bash
|
||||
# deploy.sh — One-command Deep Dive deployment
|
||||
# Issue: #830 — Sovereign NotebookLM Daily Briefing
|
||||
#
|
||||
# Usage:
|
||||
# ./deploy.sh --dry-run # Build + test only
|
||||
# ./deploy.sh --live # Build + install daily timer
|
||||
|
||||
set -euo pipefail
|
||||
|
||||
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
|
||||
COMPOSE_FILE="$SCRIPT_DIR/docker-compose.yml"
|
||||
MODE="dry-run"
|
||||
|
||||
RED='\033[0;31m'
|
||||
GREEN='\033[0;32m'
|
||||
YELLOW='\033[1;33m'
|
||||
NC='\033[0m'
|
||||
|
||||
pass() { echo -e "${GREEN}[PASS]${NC} $*"; }
|
||||
fail() { echo -e "${RED}[FAIL]${NC} $*"; }
|
||||
info() { echo -e "${YELLOW}[INFO]${NC} $*"; }
|
||||
|
||||
usage() {
|
||||
echo "Usage: $0 [--dry-run | --live]"
|
||||
echo " --dry-run Build image and run a dry-run test (default)"
|
||||
echo " --live Build image, run test, and install systemd timer"
|
||||
exit 1
|
||||
}
|
||||
|
||||
if [[ $# -gt 0 ]]; then
|
||||
case "$1" in
|
||||
--dry-run) MODE="dry-run" ;;
|
||||
--live) MODE="live" ;;
|
||||
-h|--help) usage ;;
|
||||
*) usage ;;
|
||||
esac
|
||||
fi
|
||||
|
||||
info "=================================================="
|
||||
info "Deep Dive Deployment — Issue #830"
|
||||
info "Mode: $MODE"
|
||||
info "=================================================="
|
||||
|
||||
# --- Prerequisites ---
|
||||
info "Checking prerequisites..."
|
||||
|
||||
if ! command -v docker >/dev/null 2>&1; then
|
||||
fail "Docker is not installed"
|
||||
exit 1
|
||||
fi
|
||||
pass "Docker installed"
|
||||
|
||||
if ! docker compose version >/dev/null 2>&1 && ! docker-compose version >/dev/null 2>&1; then
|
||||
fail "Docker Compose is not installed"
|
||||
exit 1
|
||||
fi
|
||||
pass "Docker Compose installed"
|
||||
|
||||
if [[ ! -f "$SCRIPT_DIR/config.yaml" ]]; then
|
||||
fail "config.yaml not found in $SCRIPT_DIR"
|
||||
info "Copy config.yaml.example or create one before deploying."
|
||||
exit 1
|
||||
fi
|
||||
pass "config.yaml exists"
|
||||
|
||||
# --- Build ---
|
||||
info "Building Deep Dive image..."
|
||||
cd "$SCRIPT_DIR"
|
||||
docker compose -f "$COMPOSE_FILE" build deepdive
|
||||
pass "Image built successfully"
|
||||
|
||||
# --- Dry-run test ---
|
||||
info "Running dry-run pipeline test..."
|
||||
docker compose -f "$COMPOSE_FILE" run --rm deepdive --dry-run --since 48
|
||||
pass "Dry-run test passed"
|
||||
|
||||
# --- Live mode: install timer ---
|
||||
if [[ "$MODE" == "live" ]]; then
|
||||
info "Installing daily execution timer..."
|
||||
|
||||
SYSTEMD_DIR="$HOME/.config/systemd/user"
|
||||
mkdir -p "$SYSTEMD_DIR"
|
||||
|
||||
# Generate a service that runs via docker compose
|
||||
cat > "$SYSTEMD_DIR/deepdive.service" <<EOF
|
||||
[Unit]
|
||||
Description=Deep Dive Daily Intelligence Briefing
|
||||
After=docker.service
|
||||
|
||||
[Service]
|
||||
Type=oneshot
|
||||
WorkingDirectory=$SCRIPT_DIR
|
||||
ExecStart=/usr/bin/docker compose -f $COMPOSE_FILE run --rm deepdive --today
|
||||
EOF
|
||||
|
||||
cat > "$SYSTEMD_DIR/deepdive.timer" <<EOF
|
||||
[Unit]
|
||||
Description=Run Deep Dive daily at 06:00
|
||||
|
||||
[Timer]
|
||||
OnCalendar=*-*-* 06:00:00
|
||||
Persistent=true
|
||||
|
||||
[Install]
|
||||
WantedBy=timers.target
|
||||
EOF
|
||||
|
||||
systemctl --user daemon-reload
|
||||
systemctl --user enable deepdive.timer
|
||||
systemctl --user start deepdive.timer || true
|
||||
|
||||
pass "Systemd timer installed and started"
|
||||
info "Check status: systemctl --user status deepdive.timer"
|
||||
|
||||
info "=================================================="
|
||||
info "Deep Dive is now deployed for live delivery!"
|
||||
info "=================================================="
|
||||
else
|
||||
info "=================================================="
|
||||
info "Deployment test successful."
|
||||
info "Run './deploy.sh --live' to enable daily automation."
|
||||
info "=================================================="
|
||||
fi
|
||||
54
intelligence/deepdive/docker-compose.yml
Normal file
54
intelligence/deepdive/docker-compose.yml
Normal file
@@ -0,0 +1,54 @@
|
||||
# Deep Dive — Full Containerized Deployment
|
||||
# Issue: #830 — Sovereign NotebookLM Daily Briefing
|
||||
#
|
||||
# Usage:
|
||||
# docker compose up -d # Start stack
|
||||
# docker compose run --rm deepdive --dry-run # Test pipeline
|
||||
# docker compose run --rm deepdive --today # Live run
|
||||
#
|
||||
# For daily automation, use systemd timer or host cron calling:
|
||||
# docker compose -f /path/to/docker-compose.yml run --rm deepdive --today
|
||||
|
||||
services:
|
||||
deepdive:
|
||||
build:
|
||||
context: .
|
||||
dockerfile: Dockerfile
|
||||
container_name: deepdive
|
||||
image: deepdive:latest
|
||||
volumes:
|
||||
# Mount your config from host
|
||||
- ./config.yaml:/app/config.yaml:ro
|
||||
# Persist cache and outputs
|
||||
- deepdive-cache:/app/cache
|
||||
- deepdive-output:/app/output
|
||||
environment:
|
||||
- OPENAI_API_KEY=${OPENAI_API_KEY:-}
|
||||
- ANTHROPIC_API_KEY=${ANTHROPIC_API_KEY:-}
|
||||
- ELEVENLABS_API_KEY=${ELEVENLABS_API_KEY:-}
|
||||
- TELEGRAM_BOT_TOKEN=${TELEGRAM_BOT_TOKEN:-}
|
||||
- TELEGRAM_HOME_CHANNEL=${TELEGRAM_HOME_CHANNEL:-}
|
||||
- DEEPDIVE_CACHE_DIR=/app/cache
|
||||
command: ["--dry-run"]
|
||||
# Optional: attach to Ollama for local LLM inference
|
||||
# networks:
|
||||
# - deepdive-net
|
||||
|
||||
# Optional: Local LLM backend (uncomment if using local inference)
|
||||
# ollama:
|
||||
# image: ollama/ollama:latest
|
||||
# container_name: deepdive-ollama
|
||||
# volumes:
|
||||
# - ollama-models:/root/.ollama
|
||||
# ports:
|
||||
# - "11434:11434"
|
||||
# networks:
|
||||
# - deepdive-net
|
||||
|
||||
volumes:
|
||||
deepdive-cache:
|
||||
deepdive-output:
|
||||
# ollama-models:
|
||||
|
||||
# networks:
|
||||
# deepdive-net:
|
||||
16
manifest.json
Normal file
16
manifest.json
Normal file
@@ -0,0 +1,16 @@
|
||||
{
|
||||
"name": "The Nexus — Timmy's Sovereign Home",
|
||||
"short_name": "The Nexus",
|
||||
"description": "A sovereign 3D world for Timmy, the local-first AI agent.",
|
||||
"start_url": "/",
|
||||
"display": "standalone",
|
||||
"background_color": "#050510",
|
||||
"theme_color": "#4af0c0",
|
||||
"icons": [
|
||||
{
|
||||
"src": "/favicon.ico",
|
||||
"sizes": "64x64",
|
||||
"type": "image/x-icon"
|
||||
}
|
||||
]
|
||||
}
|
||||
8
robots.txt
Normal file
8
robots.txt
Normal file
@@ -0,0 +1,8 @@
|
||||
User-agent: *
|
||||
Allow: /
|
||||
Disallow: /api/
|
||||
Disallow: /admin/
|
||||
Disallow: /user/
|
||||
Disallow: /explore/
|
||||
|
||||
Sitemap: https://forge.alexanderwhitestone.com/sitemap.xml
|
||||
122
server.py
122
server.py
@@ -1,37 +1,119 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
The Nexus WebSocket Gateway — Robust broadcast bridge for Timmy's consciousness.
|
||||
This server acts as the central hub for the-nexus, connecting the mind (nexus_think.py),
|
||||
the body (Evennia/Morrowind), and the visualization surface.
|
||||
"""
|
||||
import asyncio
|
||||
import websockets
|
||||
import json
|
||||
import logging
|
||||
import signal
|
||||
import sys
|
||||
from typing import Set
|
||||
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
clients = set()
|
||||
import websockets
|
||||
|
||||
async def broadcast_handler(websocket):
|
||||
# Configuration
|
||||
PORT = 8765
|
||||
HOST = "0.0.0.0" # Allow external connections if needed
|
||||
|
||||
# Logging setup
|
||||
logging.basicConfig(
|
||||
level=logging.INFO,
|
||||
format='%(asctime)s [%(levelname)s] %(message)s',
|
||||
datefmt='%Y-%m-%d %H:%M:%S'
|
||||
)
|
||||
logger = logging.getLogger("nexus-gateway")
|
||||
|
||||
# State
|
||||
clients: Set[websockets.WebSocketServerProtocol] = set()
|
||||
|
||||
async def broadcast_handler(websocket: websockets.WebSocketServerProtocol):
|
||||
"""Handles individual client connections and message broadcasting."""
|
||||
clients.add(websocket)
|
||||
logging.info(f"Client connected. Total clients: {len(clients)}")
|
||||
addr = websocket.remote_address
|
||||
logger.info(f"Client connected from {addr}. Total clients: {len(clients)}")
|
||||
|
||||
try:
|
||||
async for message in websocket:
|
||||
# Parse for logging/validation if it's JSON
|
||||
try:
|
||||
data = json.loads(message)
|
||||
msg_type = data.get("type", "unknown")
|
||||
# Optional: log specific important message types
|
||||
if msg_type in ["agent_register", "thought", "action"]:
|
||||
logger.debug(f"Received {msg_type} from {addr}")
|
||||
except (json.JSONDecodeError, TypeError):
|
||||
pass
|
||||
|
||||
# Broadcast to all OTHER clients
|
||||
if not clients:
|
||||
continue
|
||||
|
||||
disconnected = set()
|
||||
# Create broadcast tasks for efficiency
|
||||
tasks = []
|
||||
for client in clients:
|
||||
if client != websocket:
|
||||
try:
|
||||
await client.send(message)
|
||||
except Exception as e:
|
||||
logging.error(f"Failed to send to a client: {e}")
|
||||
disconnected.add(client)
|
||||
clients.difference_update(disconnected)
|
||||
if client != websocket and client.open:
|
||||
tasks.append(asyncio.create_task(client.send(message)))
|
||||
|
||||
if tasks:
|
||||
results = await asyncio.gather(*tasks, return_exceptions=True)
|
||||
for i, result in enumerate(results):
|
||||
if isinstance(result, Exception):
|
||||
# Find the client that failed
|
||||
target_client = [c for c in clients if c != websocket][i]
|
||||
logger.error(f"Failed to send to a client {target_client.remote_address}: {result}")
|
||||
disconnected.add(target_client)
|
||||
|
||||
if disconnected:
|
||||
clients.difference_update(disconnected)
|
||||
|
||||
except websockets.exceptions.ConnectionClosed:
|
||||
pass
|
||||
logger.debug(f"Connection closed by client {addr}")
|
||||
except Exception as e:
|
||||
logger.error(f"Error handling client {addr}: {e}")
|
||||
finally:
|
||||
clients.discard(websocket) # discard is safe if not present
|
||||
logging.info(f"Client disconnected. Total clients: {len(clients)}")
|
||||
clients.discard(websocket)
|
||||
logger.info(f"Client disconnected {addr}. Total clients: {len(clients)}")
|
||||
|
||||
async def main():
|
||||
port = 8765
|
||||
logging.info(f"Starting WS gateway on ws://localhost:{port}")
|
||||
async with websockets.serve(broadcast_handler, "localhost", port):
|
||||
await asyncio.Future() # Run forever
|
||||
"""Main server loop with graceful shutdown."""
|
||||
logger.info(f"Starting Nexus WS gateway on ws://{HOST}:{PORT}")
|
||||
|
||||
# Set up signal handlers for graceful shutdown
|
||||
loop = asyncio.get_running_loop()
|
||||
stop = loop.create_future()
|
||||
|
||||
def shutdown():
|
||||
if not stop.done():
|
||||
stop.set_result(None)
|
||||
|
||||
for sig in (signal.SIGINT, signal.SIGTERM):
|
||||
try:
|
||||
loop.add_signal_handler(sig, shutdown)
|
||||
except NotImplementedError:
|
||||
# Signal handlers not supported on Windows
|
||||
pass
|
||||
|
||||
async with websockets.serve(broadcast_handler, HOST, PORT):
|
||||
logger.info("Gateway is ready and listening.")
|
||||
await stop
|
||||
|
||||
logger.info("Shutting down Nexus WS gateway...")
|
||||
# Close all client connections
|
||||
if clients:
|
||||
logger.info(f"Closing {len(clients)} active connections...")
|
||||
close_tasks = [client.close() for client in clients]
|
||||
await asyncio.gather(*close_tasks, return_exceptions=True)
|
||||
|
||||
logger.info("Shutdown complete.")
|
||||
|
||||
if __name__ == "__main__":
|
||||
asyncio.run(main())
|
||||
try:
|
||||
asyncio.run(main())
|
||||
except KeyboardInterrupt:
|
||||
pass
|
||||
except Exception as e:
|
||||
logger.critical(f"Fatal server error: {e}")
|
||||
sys.exit(1)
|
||||
|
||||
Reference in New Issue
Block a user