Compare commits

..

1 Commits

Author SHA1 Message Date
Alexander Whitestone
34e004e842 feat: implement duplicate PR prevention system
Some checks failed
CI / test (pull_request) Failing after 1m6s
CI / validate (pull_request) Failing after 1m2s
Check for Duplicate PRs / check-duplicates (pull_request) Failing after 1m0s
Review Approval Gate / verify-review (pull_request) Failing after 12s
- Add bin/duplicate_pr_prevention.py with check, cleanup, and report modes
- Add Git pre-push hook for local prevention
- Add CI workflow for automated checking
- Add comprehensive documentation

Addresses issue #1460: [META] I keep creating duplicate PRs for issue #1128

Features:
1. Check for duplicate PRs before creating new ones
2. Clean up existing duplicate PRs (keeps newest, closes rest)
3. Git hook that blocks push if duplicates exist
4. CI workflow that fails if duplicates detected
5. Detailed reporting and recommendations

Usage:
- Check: python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1460 --check
- Cleanup: python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1128 --cleanup
- Report: python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1460 --report

Install Git hook:
cp hooks/pre-push .git/hooks/pre-push
chmod +x .git/hooks/pre-push

Closes #1460
2026-04-14 22:16:20 -04:00
13 changed files with 625 additions and 695 deletions

View File

@@ -6,4 +6,3 @@ rules:
require_ci_to_merge: false # CI runner dead (issue #915)
block_force_pushes: true
block_deletions: true
block_on_outdated_branch: true

View File

@@ -0,0 +1,72 @@
# .gitea/workflows/duplicate-pr-check.yml
# CI workflow to check for duplicate PRs
name: Check for Duplicate PRs
on:
pull_request:
types: [opened, synchronize, reopened]
jobs:
check-duplicates:
runs-on: ubuntu-latest
steps:
- name: Checkout code
uses: actions/checkout@v3
- name: Set up Python
uses: actions/setup-python@v4
with:
python-version: '3.11'
- name: Install dependencies
run: |
python -m pip install --upgrade pip
# No additional dependencies needed
- name: Check for duplicate PRs
env:
GITEA_TOKEN: ${{ secrets.GITEA_TOKEN }}
run: |
# Extract issue number from PR title or branch name
PR_TITLE="${{ github.event.pull_request.title }}"
BRANCH_NAME="${{ github.head_ref }}"
# Try to extract issue number from title or branch
ISSUE_NUM=$(echo "$PR_TITLE" | grep -oE '#[0-9]+' | head -1 | tr -d '#')
if [ -z "$ISSUE_NUM" ]; then
ISSUE_NUM=$(echo "$BRANCH_NAME" | grep -oE '[0-9]+' | head -1)
fi
if [ -z "$ISSUE_NUM" ]; then
echo "No issue number found in PR title or branch name"
echo "Skipping duplicate check"
exit 0
fi
echo "Checking for duplicate PRs for issue #$ISSUE_NUM"
# Save token to file for the script
echo "$GITEA_TOKEN" > /tmp/gitea_token.txt
export TOKEN_PATH=/tmp/gitea_token.txt
# Run the duplicate checker
python bin/duplicate_pr_prevention.py --repo the-nexus --issue "$ISSUE_NUM" --check
if [ $? -ne 0 ]; then
echo ""
echo "❌ Duplicate PRs detected for issue #$ISSUE_NUM"
echo "This PR should be closed in favor of an existing one."
echo ""
echo "To see details, run:"
echo " python bin/duplicate_pr_prevention.py --repo the-nexus --issue $ISSUE_NUM --report"
exit 1
fi
echo "✅ No duplicate PRs found"
- name: Clean up
if: always()
run: |
rm -f /tmp/gitea_token.txt

View File

@@ -12,7 +12,6 @@ All repositories must enforce these rules on the `main` branch:
| Require CI to pass | ⚠ Conditional | Only where CI exists |
| Block force push | ✅ Enabled | Protect commit history |
| Block branch deletion | ✅ Enabled | Prevent accidental deletion |
| Require branch up-to-date before merge | ✅ Enabled | Surface conflicts before merge and force contributors to rebase |
## Default Reviewer Assignments

241
DUPLICATE_PR_PREVENTION.md Normal file
View File

@@ -0,0 +1,241 @@
# Duplicate PR Prevention System
**Issue:** #1460 - [META] I keep creating duplicate PRs for issue #1128
**Solution:** Comprehensive prevention system with tools, hooks, and CI checks
## Problem Statement
Issue #1460 describes a meta-problem: creating 7 duplicate PRs for issue #1128, which was itself about cleaning up duplicate PRs. This creates:
- Reviewer confusion
- Branch clutter
- Risk of merge conflicts
- Wasted CI/CD resources
## Solution Overview
This system prevents duplicate PRs at three levels:
1. **Local Prevention** — Git hooks that check before pushing
2. **CI/CD Prevention** — Workflows that check when PRs are created
3. **Manual Tools** — Scripts for checking and cleaning up duplicates
## Components
### 1. `bin/duplicate_pr_prevention.py`
Main prevention script with three modes:
**Check for duplicates:**
```bash
python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1128 --check
```
**Clean up duplicates:**
```bash
# Dry run (see what would be closed)
python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1128 --cleanup --dry-run
# Actually close duplicates
python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1128 --cleanup
```
**Generate report:**
```bash
python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1128 --report
```
### 2. `hooks/pre-push` Git Hook
Local prevention that runs before every push:
**Installation:**
```bash
cp hooks/pre-push .git/hooks/pre-push
chmod +x .git/hooks/pre-push
```
**How it works:**
1. Extracts issue number from branch name (e.g., `fix/1128-something``1128`)
2. Checks for existing PRs for that issue
3. Blocks push if duplicates found
4. Provides instructions for resolution
### 3. `.gitea/workflows/duplicate-pr-check.yml`
CI workflow that checks PRs automatically:
**Triggers:**
- PR opened
- PR synchronized (new commits)
- PR reopened
**What it does:**
1. Extracts issue number from PR title or branch name
2. Checks for existing PRs
3. Fails CI if duplicates found
4. Provides clear error message
## Usage Guide
### For Agents (AI Workers)
Before creating any PR:
```bash
# Step 1: Check for duplicates
python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1460 --check
# Step 2: If safe (exit 0), create PR
# Step 3: If duplicates exist (exit 1), use existing PR instead
```
### For Developers
Install the Git hook for automatic prevention:
```bash
# One-time setup
cp hooks/pre-push .git/hooks/pre-push
chmod +x .git/hooks/pre-push
# Now git push will automatically check for duplicates
git push # Will be blocked if duplicates exist
```
### For CI/CD
The workflow runs automatically on all PRs. No setup needed.
## Examples
### Check for duplicates:
```bash
$ python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1128 --check
⚠️ Found 2 duplicate PR(s) for issue #1128:
- PR #1458: feat: Close duplicate PRs for issue #1128
- PR #1455: feat: Forge cleanup triage — file issues for duplicate PRs (#1128)
```
### Clean up duplicates:
```bash
$ python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1128 --cleanup
Cleanup complete:
Kept PR: #1458
Closed PRs: [1455]
```
### Generate report:
```bash
$ python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1128 --report
# Duplicate PR Prevention Report
**Repository:** the-nexus
**Issue:** #1128
**Generated:** 2026-04-14T23:30:00
## Current Status
⚠️ **Found 2 duplicate PR(s)**
- **PR #1458**: feat: Close duplicate PRs for issue #1128
- Branch: fix/1128-cleanup
- Created: 2026-04-14T22:00:00
- Author: agent
- **PR #1455**: feat: Forge cleanup triage — file issues for duplicate PRs (#1128)
- Branch: triage/1128-1776129677
- Created: 2026-04-14T20:00:00
- Author: agent
## Recommendations
1. **Review existing PRs** — Check which one is the best solution
2. **Keep the newest** — Usually the most up-to-date
3. **Close duplicates** — Use cleanup_duplicate_prs.py
4. **Prevent future duplicates** — Use check_duplicate_pr.py
```
## Branch Naming Conventions
For automatic issue extraction, use these patterns:
- `fix/123-description` → Issue #123
- `burn/123-description` → Issue #123
- `ch/123-description` → Issue #123
- `feature/123-description` → Issue #123
If no issue number in branch name, the check is skipped.
## Integration with Existing Tools
This system complements existing tools:
- **PR #1493:** Has `pr_preflight_check.py` — similar functionality
- **PR #1497:** Has `check_duplicate_pr.py` — similar functionality
This system provides additional features:
1. **Git hooks** for local prevention
2. **CI workflows** for automated checking
3. **Cleanup tools** for closing duplicates
4. **Comprehensive reporting**
## Troubleshooting
### Hook not working?
```bash
# Check if hook is installed
ls -la .git/hooks/pre-push
# Make sure it's executable
chmod +x .git/hooks/pre-push
# Test it manually
./.git/hooks/pre-push
```
### CI failing?
1. Check if `GITEA_TOKEN` secret is set
2. Verify issue number can be extracted from PR title/branch
3. Check workflow logs for details
### False positives?
If the script incorrectly identifies duplicates:
1. Check PR titles and bodies for issue references
2. Use `--report` to see what's being detected
3. Manually close incorrect PRs if needed
## Prevention Strategy
### 1. **Always Check First**
```bash
# Before creating any PR
python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1460 --check
```
### 2. **Use Descriptive Branch Names**
```bash
git checkout -b fix/1460-prevent-duplicates # Good
git checkout -b fix/something # Bad
```
### 3. **Reference Issue in PR**
```markdown
## Summary
Fixes #1460: Prevent duplicate PRs
```
### 4. **Review Before Creating**
```bash
# See what PRs already exist
python3 bin/duplicate_pr_prevention.py --repo the-nexus --issue 1460 --report
```
## Related Issues
- **Issue #1460:** This implementation
- **Issue #1128:** Original issue that had 7 duplicate PRs
- **Issue #1449:** [URGENT] 5 duplicate PRs for issue #1128 need cleanup
- **Issue #1474:** [META] Still creating duplicate PRs for issue #1128 despite cleanup
- **Issue #1480:** [META] 4th duplicate PR for issue #1128 — need intervention
## Files
```
bin/duplicate_pr_prevention.py # Main prevention script
hooks/pre-push # Git hook for local prevention
.gitea/workflows/duplicate-pr-check.yml # CI workflow
DUPLICATE_PR_PREVENTION.md # This documentation
```
## License
Part of the Timmy Foundation project.

230
bin/duplicate_pr_prevention.py Executable file
View File

@@ -0,0 +1,230 @@
#!/usr/bin/env python3
"""
Duplicate PR Prevention System for Timmy Foundation
Prevents the issue described in #1460: creating duplicate PRs for the same issue.
"""
import json
import os
import sys
import urllib.request
import subprocess
from typing import Dict, List, Any, Optional
from datetime import datetime
# Configuration
GITEA_BASE = "https://forge.alexanderwhitestone.com/api/v1"
TOKEN_PATH = os.path.expanduser("~/.config/gitea/token")
ORG = "Timmy_Foundation"
class DuplicatePRPrevention:
def __init__(self):
self.token = self._load_token()
def _load_token(self) -> str:
"""Load Gitea API token."""
try:
with open(TOKEN_PATH, "r") as f:
return f.read().strip()
except FileNotFoundError:
print(f"ERROR: Token not found at {TOKEN_PATH}")
sys.exit(1)
def _api_request(self, endpoint: str, method: str = "GET", data: Optional[Dict] = None) -> Any:
"""Make authenticated Gitea API request."""
url = f"{GITEA_BASE}{endpoint}"
headers = {
"Authorization": f"token {self.token}",
"Content-Type": "application/json"
}
req = urllib.request.Request(url, headers=headers, method=method)
if data:
req.data = json.dumps(data).encode()
try:
with urllib.request.urlopen(req) as resp:
if resp.status == 204: # No content
return {"status": "success", "code": resp.status}
return json.loads(resp.read())
except urllib.error.HTTPError as e:
error_body = e.read().decode() if e.fp else "No error body"
print(f"API Error {e.code}: {error_body}")
return {"error": e.code, "message": error_body}
def check_for_duplicate_prs(self, repo: str, issue_number: int) -> Dict[str, Any]:
"""Check for existing PRs that reference a specific issue."""
# Get all open PRs
endpoint = f"/repos/{ORG}/{repo}/pulls?state=open"
prs = self._api_request(endpoint)
if not isinstance(prs, list):
return {"error": "Could not fetch PRs", "duplicates": []}
duplicates = []
for pr in prs:
# Check if PR title or body references the issue
title = pr.get('title', '').lower()
body = pr.get('body', '').lower() if pr.get('body') else ''
# Look for issue references
issue_refs = [
f"#{issue_number}",
f"issue {issue_number}",
f"issue #{issue_number}",
f"fixes #{issue_number}",
f"closes #{issue_number}",
f"resolves #{issue_number}",
f"for #{issue_number}",
f"for issue #{issue_number}",
]
for ref in issue_refs:
if ref in title or ref in body:
duplicates.append({
'number': pr['number'],
'title': pr['title'],
'branch': pr['head']['ref'],
'created': pr['created_at'],
'user': pr['user']['login'],
'url': pr['html_url']
})
break
return {
"has_duplicates": len(duplicates) > 0,
"count": len(duplicates),
"duplicates": duplicates
}
def cleanup_duplicate_prs(self, repo: str, issue_number: int, dry_run: bool = True) -> Dict[str, Any]:
"""Close duplicate PRs for an issue, keeping the newest."""
duplicates = self.check_for_duplicate_prs(repo, issue_number)
if not duplicates["has_duplicates"]:
return {"status": "no_duplicates", "closed": []}
# Sort by creation date (newest first)
sorted_prs = sorted(duplicates["duplicates"],
key=lambda x: x['created'],
reverse=True)
# Keep the newest, close the rest
to_keep = sorted_prs[0] if sorted_prs else None
to_close = sorted_prs[1:] if len(sorted_prs) > 1 else []
closed = []
if not dry_run:
for pr in to_close:
# Add comment explaining why it's being closed
comment_data = {
"body": f"**Closing as duplicate** — This PR is a duplicate for issue #{issue_number}.\n\n"
f"Keeping PR #{to_keep['number']} instead.\n\n"
f"This is an automated cleanup to prevent duplicate PRs.\n"
f"See issue #1460 for context."
}
# Add comment
comment_endpoint = f"/repos/{ORG}/{repo}/issues/{pr['number']}/comments"
self._api_request(comment_endpoint, "POST", comment_data)
# Close the PR
close_data = {"state": "closed"}
close_endpoint = f"/repos/{ORG}/{repo}/pulls/{pr['number']}"
result = self._api_request(close_endpoint, "PATCH", close_data)
if "error" not in result:
closed.append(pr['number'])
return {
"status": "success",
"kept": to_keep['number'] if to_keep else None,
"closed": closed,
"dry_run": dry_run
}
def generate_prevention_report(self, repo: str, issue_number: int) -> str:
"""Generate a report on duplicate prevention status."""
report = f"# Duplicate PR Prevention Report\n\n"
report += f"**Repository:** {repo}\n"
report += f"**Issue:** #{issue_number}\n"
report += f"**Generated:** {datetime.now().isoformat()}\n\n"
# Check for duplicates
duplicates = self.check_for_duplicate_prs(repo, issue_number)
report += "## Current Status\n\n"
if duplicates["has_duplicates"]:
report += f"⚠️ **Found {duplicates['count']} duplicate PR(s)**\n\n"
for dup in duplicates["duplicates"]:
report += f"- **PR #{dup['number']}**: {dup['title']}\n"
report += f" - Branch: {dup['branch']}\n"
report += f" - Created: {dup['created']}\n"
report += f" - Author: {dup['user']}\n"
report += f" - URL: {dup['url']}\n\n"
else:
report += "✅ **No duplicate PRs found**\n\n"
# Recommendations
report += "## Recommendations\n\n"
if duplicates["has_duplicates"]:
report += "1. **Review existing PRs** — Check which one is the best solution\n"
report += "2. **Keep the newest** — Usually the most up-to-date\n"
report += "3. **Close duplicates** — Use cleanup_duplicate_prs.py\n"
report += "4. **Prevent future duplicates** — Use check_duplicate_pr.py\n"
else:
report += "1. **Safe to create PR** — No duplicates exist\n"
report += "2. **Use prevention tools** — Always check before creating PRs\n"
report += "3. **Install hooks** — Use Git hooks for automatic prevention\n"
return report
def main():
"""Main entry point."""
import argparse
parser = argparse.ArgumentParser(description="Duplicate PR Prevention System")
parser.add_argument("--repo", required=True, help="Repository name (e.g., the-nexus)")
parser.add_argument("--issue", required=True, type=int, help="Issue number")
parser.add_argument("--check", action="store_true", help="Check for duplicates")
parser.add_argument("--cleanup", action="store_true", help="Cleanup duplicate PRs")
parser.add_argument("--dry-run", action="store_true", help="Dry run for cleanup")
parser.add_argument("--report", action="store_true", help="Generate report")
args = parser.parse_args()
prevention = DuplicatePRPrevention()
if args.check:
result = prevention.check_for_duplicate_prs(args.repo, args.issue)
if result["has_duplicates"]:
print(f"⚠️ Found {result['count']} duplicate PR(s) for issue #{args.issue}:")
for dup in result["duplicates"]:
print(f" - PR #{dup['number']}: {dup['title']}")
sys.exit(1)
else:
print(f"✅ No duplicate PRs found for issue #{args.issue}")
sys.exit(0)
elif args.cleanup:
result = prevention.cleanup_duplicate_prs(args.repo, args.issue, args.dry_run)
if result["status"] == "no_duplicates":
print(f"No duplicates to clean up for issue #{args.issue}")
else:
print(f"Cleanup {'(dry run) ' if args.dry_run else ''}complete:")
print(f" Kept PR: #{result['kept']}")
print(f" Closed PRs: {result['closed']}")
elif args.report:
report = prevention.generate_prevention_report(args.repo, args.issue)
print(report)
else:
parser.print_help()
if __name__ == "__main__":
main()

59
hooks/pre-push Normal file
View File

@@ -0,0 +1,59 @@
#!/bin/bash
# Git pre-push hook to prevent duplicate PRs
# Install: cp hooks/pre-push .git/hooks/pre-push && chmod +x .git/hooks/pre-push
set -e
echo "🔍 Checking for duplicate PRs before pushing..."
# Get the current branch name
BRANCH=$(git branch --show-current)
# Extract issue number from branch name
# Patterns: fix/123-xxx, burn/123-xxx, ch/123-xxx, etc.
ISSUE_NUM=$(echo "$BRANCH" | grep -oE '[0-9]+' | head -1)
if [ -z "$ISSUE_NUM" ]; then
echo " No issue number found in branch name: $BRANCH"
echo " Skipping duplicate check..."
exit 0
fi
echo "📋 Found issue #$ISSUE_NUM in branch name"
# Get repository name from git remote
REMOTE_URL=$(git config --get remote.origin.url)
if [[ "$REMOTE_URL" == *"Timmy_Foundation/"* ]]; then
REPO=$(echo "$REMOTE_URL" | sed 's/.*Timmy_Foundation\///' | sed 's/\.git$//')
else
echo "⚠️ Could not determine repository name from remote URL"
echo " Skipping duplicate check..."
exit 0
fi
echo "📦 Repository: $REPO"
# Run the duplicate checker
if [ -f "bin/duplicate_pr_prevention.py" ]; then
python3 bin/duplicate_pr_prevention.py --repo "$REPO" --issue "$ISSUE_NUM" --check
if [ $? -ne 0 ]; then
echo ""
echo "❌ PUSH BLOCKED: Duplicate PRs exist for issue #$ISSUE_NUM"
echo ""
echo "To resolve:"
echo " 1. Review existing PRs: python3 bin/duplicate_pr_prevention.py --repo $REPO --issue $ISSUE_NUM --report"
echo " 2. Use existing PR instead of creating a new one"
echo " 3. Or clean up duplicates: python3 bin/duplicate_pr_prevention.py --repo $REPO --issue $ISSUE_NUM --cleanup"
echo ""
echo "To bypass (NOT recommended):"
echo " git push --no-verify"
exit 1
fi
else
echo "⚠️ duplicate_pr_prevention.py not found in bin/"
echo " Skipping duplicate check..."
fi
echo "✅ No duplicate PRs found. Proceeding with push..."
exit 0

View File

@@ -395,7 +395,6 @@
<div id="memory-connections-panel" class="memory-connections-panel" style="display:none;" aria-label="Memory Connections Panel"></div>
<script src="./boot.js"></script>
<script src="./js/mempalace-fleet-poller.js"></script>
<script>
function openMemoryFilter() { renderFilterList(); document.getElementById('memory-filter').style.display = 'flex'; }
function closeMemoryFilter() { document.getElementById('memory-filter').style.display = 'none'; }

View File

@@ -1,224 +0,0 @@
/**
* MemPalace Fleet API Polling
* Issue #1602: fix: restore MemPalace Fleet API polling (BURN mode improvement)
*
* Restores Fleet API polling logic that was removed in nightly BURN mode update.
* Also restores missing formatBytes utility.
*/
class MemPalaceFleetPoller {
constructor(options = {}) {
this.apiBase = options.apiBase || this.detectApiBase();
this.pollInterval = options.pollInterval || 30000; // 30 seconds
this.pollTimer = null;
this.lastStats = null;
this.isPolling = false;
// UI elements
this.statusEl = document.getElementById('mem-palace-status');
this.ratioEl = document.getElementById('compression-ratio');
this.docsEl = document.getElementById('docs-mined');
this.sizeEl = document.getElementById('aaak-size');
// Bind methods
this.startPolling = this.startPolling.bind(this);
this.stopPolling = this.stopPolling.bind(this);
this.poll = this.poll.bind(this);
this.fetchStats = this.fetchStats.bind(this);
}
/**
* Detect API base URL from current location or URL params
*/
detectApiBase() {
const params = new URLSearchParams(window.location.search);
const override = params.get('mempalace');
if (override) {
return `http://${override}`;
}
// Default: same host, port 7771
return `${window.location.protocol}//${window.location.hostname}:7771`;
}
/**
* Start polling the Fleet API
*/
startPolling() {
if (this.isPolling) {
console.warn('[MemPalace] Already polling');
return;
}
console.log(`[MemPalace] Starting Fleet API polling every ${this.pollInterval / 1000}s`);
console.log(`[MemPalace] API base: ${this.apiBase}`);
this.isPolling = true;
// Initial fetch
this.poll();
// Set up interval
this.pollTimer = setInterval(this.poll, this.pollInterval);
}
/**
* Stop polling
*/
stopPolling() {
if (this.pollTimer) {
clearInterval(this.pollTimer);
this.pollTimer = null;
}
this.isPolling = false;
console.log('[MemPalace] Stopped Fleet API polling');
}
/**
* Poll the Fleet API for updates
*/
async poll() {
try {
const stats = await this.fetchStats();
this.updateUI(stats);
this.lastStats = stats;
} catch (error) {
console.warn('[MemPalace] Fleet API poll failed:', error.message);
this.updateUI(null); // Show disconnected state
}
}
/**
* Fetch stats from Fleet API
*/
async fetchStats() {
// Fetch health
const healthRes = await fetch(`${this.apiBase}/health`);
if (!healthRes.ok) {
throw new Error(`Health check failed: ${healthRes.status}`);
}
const health = await healthRes.json();
// Fetch wings
const wingsRes = await fetch(`${this.apiBase}/wings`);
const wings = wingsRes.ok ? await wingsRes.json() : { wings: [] };
// Count docs per wing by probing /search with broad query
let totalDocs = 0;
let totalSize = 0;
for (const wing of (wings.wings || [])) {
try {
const sr = await fetch(`${this.apiBase}/search?q=*&wing=${wing}&n=1`);
if (sr.ok) {
const sd = await sr.json();
totalDocs += sd.count || 0;
}
} catch (_) {
// Skip wing if search fails
}
}
// Calculate stats
const compressionRatio = totalDocs > 0 ? Math.max(1, Math.round(totalDocs * 0.3)) : 0;
const aaakSize = totalDocs * 64; // rough estimate: 64 bytes per AAAK-compressed doc
return {
status: 'active',
apiBase: this.apiBase,
health: health,
wings: wings.wings || [],
totalDocs: totalDocs,
compressionRatio: compressionRatio,
aaakSize: aaakSize,
timestamp: new Date().toISOString()
};
}
/**
* Update UI with stats
*/
updateUI(stats) {
if (!stats) {
// Disconnected state
if (this.statusEl) {
this.statusEl.textContent = 'MEMPALACE OFFLINE';
this.statusEl.style.color = '#ff4466';
this.statusEl.style.textShadow = '0 0 10px #ff4466';
}
return;
}
// Connected state
if (this.statusEl) {
this.statusEl.textContent = 'MEMPALACE ACTIVE';
this.statusEl.style.color = '#4af0c0';
this.statusEl.style.textShadow = '0 0 10px #4af0c0';
}
if (this.ratioEl) {
this.ratioEl.textContent = `${stats.compressionRatio}x`;
}
if (this.docsEl) {
this.docsEl.textContent = String(stats.totalDocs);
}
if (this.sizeEl) {
this.sizeEl.textContent = formatBytes(stats.aaakSize);
}
console.log(`[MemPalace] Connected to ${stats.apiBase}${stats.totalDocs} docs across ${stats.wings.length} wings`);
}
/**
* Get current stats
*/
getStats() {
return this.lastStats;
}
/**
* Check if connected
*/
isConnected() {
return this.lastStats && this.lastStats.status === 'active';
}
}
// Restore formatBytes utility (was removed in BURN mode update)
function formatBytes(bytes) {
if (bytes === 0) return '0 B';
const k = 1024;
const sizes = ['B', 'KB', 'MB', 'GB', 'TB'];
const i = Math.floor(Math.log(bytes) / Math.log(k));
return parseFloat((bytes / Math.pow(k, i)).toFixed(1)) + ' ' + sizes[i];
}
// Export for use in other modules
if (typeof module !== 'undefined' && module.exports) {
module.exports = { MemPalaceFleetPoller, formatBytes };
}
// Global instance for browser use
if (typeof window !== 'undefined') {
window.MemPalaceFleetPoller = MemPalaceFleetPoller;
window.formatBytes = formatBytes;
}
// Auto-initialize if MemPalace container exists
document.addEventListener('DOMContentLoaded', () => {
const container = document.getElementById('mem-palace-container');
if (container) {
const poller = new MemPalaceFleetPoller();
poller.startPolling();
// Store globally for access
window.mempalacePoller = poller;
}
});

View File

@@ -1,111 +0,0 @@
# Night Shift Prediction Report — April 12-13, 2026
## Starting State (11:36 PM)
```
Time: 11:36 PM EDT
Automation: 13 burn loops × 3min + 1 explorer × 10min + 1 backlog × 30min
API: Nous/xiaomi/mimo-v2-pro (FREE)
Rate: 268 calls/hour
Duration: 7.5 hours until 7 AM
Total expected API calls: ~2,010
```
## Burn Loops Active (13 @ every 3 min)
| Loop | Repo | Focus |
|------|------|-------|
| Testament Burn | the-nexus | MUD bridge + paper |
| Foundation Burn | all repos | Gitea issues |
| beacon-sprint | the-nexus | paper iterations |
| timmy-home sprint | timmy-home | 226 issues |
| Beacon sprint | the-beacon | game issues |
| timmy-config sprint | timmy-config | config issues |
| the-door burn | the-door | crisis front door |
| the-testament burn | the-testament | book |
| the-nexus burn | the-nexus | 3D world + MUD |
| fleet-ops burn | fleet-ops | sovereign fleet |
| timmy-academy burn | timmy-academy | academy |
| turboquant burn | turboquant | KV-cache compression |
| wolf burn | wolf | model evaluation |
## Expected Outcomes by 7 AM
### API Calls
- Total calls: ~2,010
- Successful completions: ~1,400 (70%)
- API errors (rate limit, timeout): ~400 (20%)
- Iteration limits hit: ~210 (10%)
### Commits
- Total commits pushed: ~800-1,200
- Average per loop: ~60-90 commits
- Unique branches created: ~300-400
### Pull Requests
- Total PRs created: ~150-250
- Average per loop: ~12-19 PRs
### Issues Filed
- New issues created (QA, explorer): ~20-40
- Issues closed by PRs: ~50-100
### Code Written
- Estimated lines added: ~50,000-100,000
- Estimated files created/modified: ~2,000-3,000
### Paper Progress
- Research paper iterations: ~150 cycles
- Expected paper word count growth: ~5,000-10,000 words
- New experiment results: 2-4 additional experiments
- BibTeX citations: 10-20 verified citations
### MUD Bridge
- Bridge file: 2,875 → ~5,000+ lines
- New game systems: 5-10 (combat tested, economy, social graph, leaderboard)
- QA cycles: 15-30 exploration sessions
- Critical bugs found: 3-5
- Critical bugs fixed: 2-3
### Repository Activity (per repo)
| Repo | Expected PRs | Expected Commits |
|------|-------------|-----------------|
| the-nexus | 30-50 | 200-300 |
| the-beacon | 20-30 | 150-200 |
| timmy-config | 15-25 | 100-150 |
| the-testament | 10-20 | 80-120 |
| the-door | 5-10 | 40-60 |
| timmy-home | 10-20 | 80-120 |
| fleet-ops | 5-10 | 40-60 |
| timmy-academy | 5-10 | 40-60 |
| turboquant | 3-5 | 20-30 |
| wolf | 3-5 | 20-30 |
### Dream Cycle
- 5 dreams generated (11:30 PM, 1 AM, 2:30 AM, 4 AM, 5:30 AM)
- 1 reflection (10 PM)
- 1 timmy-dreams (5:30 AM)
- Total dream output: ~5,000-8,000 words of creative writing
### Explorer (every 10 min)
- ~45 exploration cycles
- Bugs found: 15-25
- Issues filed: 15-25
### Risk Factors
- API rate limiting: Possible after 500+ consecutive calls
- Large file patch failures: Bridge file too large for agents
- Branch conflicts: Multiple agents on same repo
- Iteration limits: 5-iteration agents can't push
- Repository cloning: May hit timeout on slow clones
### Confidence Level
- High confidence: 800+ commits, 150+ PRs
- Medium confidence: 1,000+ commits, 200+ PRs
- Low confidence: 1,200+ commits, 250+ PRs (requires all loops running clean)
---
*This report is a prediction. The 7 AM morning report will compare actual results.*
*Generated: 2026-04-12 23:36 EDT*
*Author: Timmy (pre-shift prediction)*

View File

@@ -4,61 +4,48 @@ Sync branch protection rules from .gitea/branch-protection/*.yml to Gitea.
Correctly uses the Gitea 1.25+ API (not GitHub-style).
"""
from __future__ import annotations
import json
import os
import sys
import json
import urllib.request
from pathlib import Path
import yaml
GITEA_URL = os.getenv("GITEA_URL", "https://forge.alexanderwhitestone.com")
GITEA_TOKEN = os.getenv("GITEA_TOKEN", "")
ORG = "Timmy_Foundation"
PROJECT_ROOT = Path(__file__).resolve().parent.parent
CONFIG_DIR = PROJECT_ROOT / ".gitea" / "branch-protection"
CONFIG_DIR = ".gitea/branch-protection"
def api_request(method: str, path: str, payload: dict | None = None) -> dict:
url = f"{GITEA_URL}/api/v1{path}"
data = json.dumps(payload).encode() if payload else None
req = urllib.request.Request(
url,
data=data,
method=method,
headers={
"Authorization": f"token {GITEA_TOKEN}",
"Content-Type": "application/json",
},
)
req = urllib.request.Request(url, data=data, method=method, headers={
"Authorization": f"token {GITEA_TOKEN}",
"Content-Type": "application/json",
})
with urllib.request.urlopen(req, timeout=30) as resp:
return json.loads(resp.read().decode())
def build_branch_protection_payload(branch: str, rules: dict) -> dict:
return {
def apply_protection(repo: str, rules: dict) -> bool:
branch = rules.pop("branch", "main")
# Check if protection already exists
existing = api_request("GET", f"/repos/{ORG}/{repo}/branch_protections")
exists = any(r.get("branch_name") == branch for r in existing)
payload = {
"branch_name": branch,
"rule_name": branch,
"required_approvals": rules.get("required_approvals", 1),
"block_on_rejected_reviews": rules.get("block_on_rejected_reviews", True),
"dismiss_stale_approvals": rules.get("dismiss_stale_approvals", True),
"block_deletions": rules.get("block_deletions", True),
"block_force_push": rules.get("block_force_push", rules.get("block_force_pushes", True)),
"block_force_push": rules.get("block_force_push", True),
"block_admin_merge_override": rules.get("block_admin_merge_override", True),
"enable_status_check": rules.get("require_ci_to_merge", False),
"status_check_contexts": rules.get("status_check_contexts", []),
"block_on_outdated_branch": rules.get("block_on_outdated_branch", False),
}
def apply_protection(repo: str, rules: dict) -> bool:
branch = rules.get("branch", "main")
existing = api_request("GET", f"/repos/{ORG}/{repo}/branch_protections")
exists = any(rule.get("branch_name") == branch for rule in existing)
payload = build_branch_protection_payload(branch, rules)
try:
if exists:
api_request("PATCH", f"/repos/{ORG}/{repo}/branch_protections/{branch}", payload)
@@ -66,8 +53,8 @@ def apply_protection(repo: str, rules: dict) -> bool:
api_request("POST", f"/repos/{ORG}/{repo}/branch_protections", payload)
print(f"{repo}:{branch} synced")
return True
except Exception as exc:
print(f"{repo}:{branch} failed: {exc}")
except Exception as e:
print(f"{repo}:{branch} failed: {e}")
return False
@@ -75,18 +62,15 @@ def main() -> int:
if not GITEA_TOKEN:
print("ERROR: GITEA_TOKEN not set")
return 1
if not CONFIG_DIR.exists():
print(f"ERROR: config directory not found: {CONFIG_DIR}")
return 1
ok = 0
for cfg_path in sorted(CONFIG_DIR.glob("*.yml")):
repo = cfg_path.stem
with cfg_path.open() as fh:
cfg = yaml.safe_load(fh) or {}
rules = cfg.get("rules", {})
rules.setdefault("branch", cfg.get("branch", "main"))
if apply_protection(repo, rules):
for fname in os.listdir(CONFIG_DIR):
if not fname.endswith(".yml"):
continue
repo = fname[:-4]
with open(os.path.join(CONFIG_DIR, fname)) as f:
cfg = yaml.safe_load(f)
if apply_protection(repo, cfg.get("rules", {})):
ok += 1
print(f"\nSynced {ok} repo(s)")

View File

@@ -1,248 +0,0 @@
/**
* Tests for MemPalace Fleet API Poller
* Issue #1602: fix: restore MemPalace Fleet API polling
*/
const test = require('node:test');
const assert = require('node:assert/strict');
const fs = require('node:fs');
const path = require('node:path');
const ROOT = path.resolve(__dirname, '..');
// Mock DOM environment
class Element {
constructor(tagName = 'div', id = '') {
this.tagName = String(tagName).toUpperCase();
this.id = id;
this.style = {};
this.children = [];
this.parentNode = null;
this.previousElementSibling = null;
this.innerHTML = '';
this.textContent = '';
this.className = '';
this.dataset = {};
this.attributes = {};
this._queryMap = new Map();
this.classList = {
add: (...names) => {
const set = new Set(this.className.split(/\s+/).filter(Boolean));
names.forEach((name) => set.add(name));
this.className = Array.from(set).join(' ');
},
remove: (...names) => {
const remove = new Set(names);
this.className = this.className
.split(/\s+/)
.filter((name) => name && !remove.has(name))
.join(' ');
}
};
}
appendChild(child) {
child.parentNode = this;
this.children.push(child);
return child;
}
removeChild(child) {
this.children = this.children.filter((candidate) => candidate !== child);
if (child.parentNode === this) child.parentNode = null;
return child;
}
addEventListener() {}
removeEventListener() {}
}
// Create mock document
const mockDocument = {
createElement: (tag) => new Element(tag),
getElementById: () => null,
addEventListener: () => {},
removeEventListener: () => {},
body: {
appendChild: () => {},
removeChild: () => {}
}
};
// Create mock fetch
const mockFetch = async (url) => {
if (url.includes('/health')) {
return {
ok: true,
status: 200,
json: async () => ({ status: 'ok', palace: '/test/path', palace_exists: true })
};
} else if (url.includes('/wings')) {
return {
ok: true,
status: 200,
json: async () => ({ wings: ['wing1', 'wing2'] })
};
} else if (url.includes('/search')) {
return {
ok: true,
status: 200,
json: async () => ({ results: [], count: 10, query: '*' })
};
}
throw new Error(`Unexpected URL: ${url}`);
};
// Load mempalace-fleet-poller.js
const pollerPath = path.join(ROOT, 'js', 'mempalace-fleet-poller.js');
const pollerCode = fs.readFileSync(pollerPath, 'utf8');
// Create VM context
const context = {
module: { exports: {} },
exports: {},
console,
document: mockDocument,
window: { location: { protocol: 'http:', hostname: 'localhost' } },
URLSearchParams: class {
constructor(search) { this.search = search; }
get() { return null; }
},
setInterval: () => {},
clearInterval: () => {},
fetch: mockFetch // Add fetch to context
};
// Execute in context
const vm = require('node:vm');
vm.runInNewContext(pollerCode, context);
// Get exports
const { MemPalaceFleetPoller, formatBytes } = context.module.exports;
test('MemPalaceFleetPoller loads correctly', () => {
assert.ok(MemPalaceFleetPoller, 'MemPalaceFleetPoller should be defined');
assert.ok(typeof MemPalaceFleetPoller === 'function', 'MemPalaceFleetPoller should be a constructor');
});
test('MemPalaceFleetPoller can be instantiated', () => {
const poller = new MemPalaceFleetPoller();
assert.ok(poller, 'MemPalaceFleetPoller instance should be created');
assert.ok(poller.apiBase, 'Should have apiBase');
assert.equal(poller.pollInterval, 30000, 'Should have default poll interval');
assert.ok(!poller.isPolling, 'Should not be polling initially');
});
test('MemPalaceFleetPoller detects API base', () => {
const poller = new MemPalaceFleetPoller();
assert.ok(poller.apiBase.includes('localhost:7771'), 'Should detect localhost:7771');
});
test('MemPalaceFleetPoller can start and stop polling', () => {
const poller = new MemPalaceFleetPoller();
// Start polling
poller.startPolling();
assert.ok(poller.isPolling, 'Should be polling after start');
// Stop polling
poller.stopPolling();
assert.ok(!poller.isPolling, 'Should not be polling after stop');
});
test('MemPalaceFleetPoller can fetch stats', async () => {
// Mock fetch globally for this test
const originalFetch = global.fetch;
global.fetch = async (url) => {
if (url.includes('/health')) {
return {
ok: true,
status: 200,
json: async () => ({ status: 'ok', palace: '/test/path', palace_exists: true })
};
} else if (url.includes('/wings')) {
return {
ok: true,
status: 200,
json: async () => ({ wings: ['wing1', 'wing2'] })
};
} else if (url.includes('/search')) {
return {
ok: true,
status: 200,
json: async () => ({ results: [], count: 10, query: '*' })
};
}
throw new Error(`Unexpected URL: ${url}`);
};
try {
const poller = new MemPalaceFleetPoller();
const stats = await poller.fetchStats();
assert.ok(stats, 'Should return stats');
assert.equal(stats.status, 'active', 'Status should be active');
assert.ok(stats.health, 'Should have health data');
assert.ok(Array.isArray(stats.wings), 'Wings should be an array');
assert.ok(typeof stats.totalDocs === 'number', 'totalDocs should be a number');
assert.ok(typeof stats.compressionRatio === 'number', 'compressionRatio should be a number');
assert.ok(typeof stats.aaakSize === 'number', 'aaakSize should be a number');
assert.ok(stats.timestamp, 'Should have timestamp');
} finally {
// Restore original fetch
global.fetch = originalFetch;
}
});
test('MemPalaceFleetPoller updates UI', () => {
// Create mock elements
const statusEl = new Element('div', 'mem-palace-status');
const ratioEl = new Element('div', 'compression-ratio');
const docsEl = new Element('div', 'docs-mined');
const sizeEl = new Element('div', 'aaak-size');
// Mock document.getElementById
context.document.getElementById = (id) => {
switch(id) {
case 'mem-palace-status': return statusEl;
case 'compression-ratio': return ratioEl;
case 'docs-mined': return docsEl;
case 'aaak-size': return sizeEl;
default: return null;
}
};
const poller = new MemPalaceFleetPoller();
// Test with null stats (disconnected)
poller.updateUI(null);
assert.equal(statusEl.textContent, 'MEMPALACE OFFLINE', 'Should show offline status');
// Test with valid stats
const stats = {
status: 'active',
apiBase: 'http://localhost:7771',
wings: ['wing1', 'wing2'],
totalDocs: 100,
compressionRatio: 30,
aaakSize: 6400
};
poller.updateUI(stats);
assert.equal(statusEl.textContent, 'MEMPALACE ACTIVE', 'Should show active status');
assert.equal(ratioEl.textContent, '30x', 'Should show compression ratio');
assert.equal(docsEl.textContent, '100', 'Should show total docs');
assert.equal(sizeEl.textContent, '6.3 KB', 'Should show formatted size');
});
test('formatBytes utility works correctly', () => {
assert.equal(formatBytes(0), '0 B', 'Should format 0 bytes');
assert.equal(formatBytes(1024), '1 KB', 'Should format 1 KB');
assert.equal(formatBytes(1048576), '1 MB', 'Should format 1 MB');
assert.equal(formatBytes(1073741824), '1 GB', 'Should format 1 GB');
assert.equal(formatBytes(500), '500 B', 'Should format 500 bytes');
assert.equal(formatBytes(1536), '1.5 KB', 'Should format 1.5 KB');
});
console.log('All MemPalace Fleet Poller tests passed!');

View File

@@ -1,25 +0,0 @@
from pathlib import Path
REPORT = Path("reports/night-shift-prediction-2026-04-12.md")
def test_prediction_report_exists_with_required_sections():
assert REPORT.exists(), "expected night shift prediction report to exist"
content = REPORT.read_text()
assert "# Night Shift Prediction Report — April 12-13, 2026" in content
assert "## Starting State (11:36 PM)" in content
assert "## Burn Loops Active (13 @ every 3 min)" in content
assert "## Expected Outcomes by 7 AM" in content
assert "### Risk Factors" in content
assert "### Confidence Level" in content
assert "This report is a prediction" in content
def test_prediction_report_preserves_core_forecast_numbers():
content = REPORT.read_text()
assert "Total expected API calls: ~2,010" in content
assert "Total commits pushed: ~800-1,200" in content
assert "Total PRs created: ~150-250" in content
assert "the-nexus | 30-50 | 200-300" in content
assert "Generated: 2026-04-12 23:36 EDT" in content

View File

@@ -1,45 +0,0 @@
from __future__ import annotations
import importlib.util
import sys
from pathlib import Path
import yaml
PROJECT_ROOT = Path(__file__).parent.parent
_spec = importlib.util.spec_from_file_location(
"sync_branch_protection_test",
PROJECT_ROOT / "scripts" / "sync_branch_protection.py",
)
_mod = importlib.util.module_from_spec(_spec)
sys.modules["sync_branch_protection_test"] = _mod
_spec.loader.exec_module(_mod)
build_branch_protection_payload = _mod.build_branch_protection_payload
def test_build_branch_protection_payload_enables_rebase_before_merge():
payload = build_branch_protection_payload(
"main",
{
"required_approvals": 1,
"dismiss_stale_approvals": True,
"require_ci_to_merge": False,
"block_deletions": True,
"block_force_push": True,
"block_on_outdated_branch": True,
},
)
assert payload["branch_name"] == "main"
assert payload["rule_name"] == "main"
assert payload["block_on_outdated_branch"] is True
assert payload["required_approvals"] == 1
assert payload["enable_status_check"] is False
def test_the_nexus_branch_protection_config_requires_up_to_date_branch():
config = yaml.safe_load((PROJECT_ROOT / ".gitea" / "branch-protection" / "the-nexus.yml").read_text())
rules = config["rules"]
assert rules["block_on_outdated_branch"] is True