Compare commits
4 Commits
fix/1505-w
...
fix/1339
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
ee2be0427c | ||
|
|
5fb8c0c513 | ||
|
|
a796453766 | ||
|
|
b4b029d2a6 |
@@ -6,4 +6,3 @@ rules:
|
||||
require_ci_to_merge: false # CI runner dead (issue #915)
|
||||
block_force_pushes: true
|
||||
block_deletions: true
|
||||
block_on_outdated_branch: true
|
||||
|
||||
1
.github/BRANCH_PROTECTION.md
vendored
1
.github/BRANCH_PROTECTION.md
vendored
@@ -12,7 +12,6 @@ All repositories must enforce these rules on the `main` branch:
|
||||
| Require CI to pass | ⚠ Conditional | Only where CI exists |
|
||||
| Block force push | ✅ Enabled | Protect commit history |
|
||||
| Block branch deletion | ✅ Enabled | Prevent accidental deletion |
|
||||
| Require branch up-to-date before merge | ✅ Enabled | Surface conflicts before merge and force contributors to rebase |
|
||||
|
||||
## Default Reviewer Assignments
|
||||
|
||||
|
||||
9
Dockerfile.preview
Normal file
9
Dockerfile.preview
Normal file
@@ -0,0 +1,9 @@
|
||||
FROM nginx:alpine
|
||||
|
||||
RUN rm /etc/nginx/conf.d/default.conf
|
||||
COPY preview/nginx.conf /etc/nginx/conf.d/default.conf
|
||||
|
||||
COPY *.html *.js *.mjs *.json *.css /usr/share/nginx/html/
|
||||
COPY nexus/ /usr/share/nginx/html/nexus/
|
||||
|
||||
EXPOSE 3000
|
||||
20
app.js
20
app.js
@@ -714,10 +714,6 @@ async function init() {
|
||||
camera = new THREE.PerspectiveCamera(65, window.innerWidth / window.innerHeight, 0.1, 1000);
|
||||
camera.position.copy(playerPos);
|
||||
|
||||
// Initialize avatar and LOD systems
|
||||
if (window.AvatarCustomization) window.AvatarCustomization.init(scene, camera);
|
||||
if (window.LODSystem) window.LODSystem.init(scene, camera);
|
||||
|
||||
updateLoad(20);
|
||||
|
||||
createSkybox();
|
||||
@@ -1253,10 +1249,16 @@ async function updateSovereignHealth() {
|
||||
const container = document.getElementById('sovereign-health-content');
|
||||
if (!container) return;
|
||||
|
||||
const params = new URLSearchParams(window.location.search);
|
||||
const metricsOverride = params.get('metrics');
|
||||
const metricsUrl = metricsOverride || `${window.location.protocol}//${window.location.host}/metrics`;
|
||||
const protocol = window.location.protocol === 'https:' ? 'wss:' : 'ws:';
|
||||
const wsStatusUrl = `${protocol}//${window.location.host}/api/world/ws`;
|
||||
|
||||
let metrics = { sovereignty_score: 100, local_sessions: 0, total_sessions: 0 };
|
||||
let daemonReachable = false;
|
||||
try {
|
||||
const res = await fetch('http://localhost:8082/metrics');
|
||||
const res = await fetch(metricsUrl);
|
||||
if (res.ok) {
|
||||
metrics = await res.json();
|
||||
daemonReachable = true;
|
||||
@@ -1269,8 +1271,8 @@ async function updateSovereignHealth() {
|
||||
{ name: 'LOCAL DAEMON', status: daemonReachable ? 'ONLINE' : 'OFFLINE' },
|
||||
{ name: 'FORGE / GITEA', url: 'https://forge.alexanderwhitestone.com', status: 'ONLINE' },
|
||||
{ name: 'NEXUS CORE', url: 'https://forge.alexanderwhitestone.com/Timmy_Foundation/the-nexus', status: 'ONLINE' },
|
||||
{ name: 'HERMES WS', url: 'ws://143.198.27.163:8765', status: wsConnected ? 'ONLINE' : 'OFFLINE' },
|
||||
{ name: 'SOVEREIGNTY', url: 'http://localhost:8082/metrics', status: metrics.sovereignty_score + '%' }
|
||||
{ name: 'HERMES WS', url: wsStatusUrl, status: wsConnected ? 'ONLINE' : 'OFFLINE' },
|
||||
{ name: 'SOVEREIGNTY', url: metricsUrl, status: metrics.sovereignty_score + '%' }
|
||||
];
|
||||
|
||||
container.innerHTML = '';
|
||||
@@ -3561,10 +3563,6 @@ function gameLoop() {
|
||||
|
||||
if (composer) { composer.render(); } else { renderer.render(scene, camera); }
|
||||
|
||||
// Update avatar and LOD systems
|
||||
if (window.AvatarCustomization && playerPos) window.AvatarCustomization.update(playerPos);
|
||||
if (window.LODSystem && playerPos) window.LODSystem.update(playerPos);
|
||||
|
||||
updateAshStorm(delta, elapsed);
|
||||
|
||||
// Project Mnemosyne - Memory Orb Animation
|
||||
|
||||
@@ -1,5 +1,3 @@
|
||||
version: "3.9"
|
||||
|
||||
services:
|
||||
nexus-main:
|
||||
build: .
|
||||
@@ -7,9 +5,21 @@ services:
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "8765:8765"
|
||||
|
||||
nexus-staging:
|
||||
build: .
|
||||
container_name: nexus-staging
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "8766:8765"
|
||||
- "8766:8765"
|
||||
|
||||
nexus-preview:
|
||||
build:
|
||||
context: .
|
||||
dockerfile: Dockerfile.preview
|
||||
container_name: nexus-preview
|
||||
restart: unless-stopped
|
||||
ports:
|
||||
- "3000:3000"
|
||||
depends_on:
|
||||
- nexus-main
|
||||
25
docs/preview-deploy.md
Normal file
25
docs/preview-deploy.md
Normal file
@@ -0,0 +1,25 @@
|
||||
# Nexus preview deploy
|
||||
|
||||
The Nexus frontend must be served over HTTP for ES modules to boot. This repo now includes a preview stack that serves the frontend on a proper URL and proxies `/api/world/ws` back to the existing Nexus WebSocket gateway.
|
||||
|
||||
## Quick start
|
||||
|
||||
```bash
|
||||
docker compose up -d nexus-main nexus-preview
|
||||
```
|
||||
|
||||
Open:
|
||||
- `http://localhost:3000`
|
||||
|
||||
The preview service serves the static frontend and proxies WebSocket traffic at:
|
||||
- `/api/world/ws`
|
||||
|
||||
## Remote preview
|
||||
|
||||
If you run the same compose stack on a VPS, the preview URL is:
|
||||
- `http://<host>:3000`
|
||||
|
||||
## Notes
|
||||
- `nexus-main` keeps serving the backend WebSocket gateway on port `8765`
|
||||
- `nexus-preview` serves the frontend on port `3000`
|
||||
- The browser can stay on a single origin because nginx proxies the WebSocket path
|
||||
@@ -395,8 +395,6 @@
|
||||
<div id="memory-connections-panel" class="memory-connections-panel" style="display:none;" aria-label="Memory Connections Panel"></div>
|
||||
|
||||
<script src="./boot.js"></script>
|
||||
<script src="./avatar-customization.js"></script>
|
||||
<script src="./lod-system.js"></script>
|
||||
<script>
|
||||
function openMemoryFilter() { renderFilterList(); document.getElementById('memory-filter').style.display = 'flex'; }
|
||||
function closeMemoryFilter() { document.getElementById('memory-filter').style.display = 'none'; }
|
||||
|
||||
186
lod-system.js
186
lod-system.js
@@ -1,186 +0,0 @@
|
||||
/**
|
||||
* LOD (Level of Detail) System for The Nexus
|
||||
*
|
||||
* Optimizes rendering when many avatars/users are visible:
|
||||
* - Distance-based LOD: far users become billboard sprites
|
||||
* - Occlusion: skip rendering users behind walls
|
||||
* - Budget: maintain 60 FPS target with 50+ avatars
|
||||
*
|
||||
* Usage:
|
||||
* LODSystem.init(scene, camera);
|
||||
* LODSystem.registerAvatar(avatarMesh, userId);
|
||||
* LODSystem.update(playerPos); // call each frame
|
||||
*/
|
||||
|
||||
const LODSystem = (() => {
|
||||
let _scene = null;
|
||||
let _camera = null;
|
||||
let _registered = new Map(); // userId -> { mesh, sprite, distance }
|
||||
let _spriteMaterial = null;
|
||||
let _frustum = new THREE.Frustum();
|
||||
let _projScreenMatrix = new THREE.Matrix4();
|
||||
|
||||
// Thresholds
|
||||
const LOD_NEAR = 15; // Full mesh within 15 units
|
||||
const LOD_FAR = 40; // Billboard beyond 40 units
|
||||
const LOD_CULL = 80; // Don't render beyond 80 units
|
||||
const SPRITE_SIZE = 1.2;
|
||||
|
||||
function init(sceneRef, cameraRef) {
|
||||
_scene = sceneRef;
|
||||
_camera = cameraRef;
|
||||
|
||||
// Create shared sprite material
|
||||
const canvas = document.createElement('canvas');
|
||||
canvas.width = 64;
|
||||
canvas.height = 64;
|
||||
const ctx = canvas.getContext('2d');
|
||||
// Simple avatar indicator: colored circle
|
||||
ctx.fillStyle = '#00ffcc';
|
||||
ctx.beginPath();
|
||||
ctx.arc(32, 32, 20, 0, Math.PI * 2);
|
||||
ctx.fill();
|
||||
ctx.fillStyle = '#0a0f1a';
|
||||
ctx.beginPath();
|
||||
ctx.arc(32, 28, 8, 0, Math.PI * 2); // head
|
||||
ctx.fill();
|
||||
|
||||
const texture = new THREE.CanvasTexture(canvas);
|
||||
_spriteMaterial = new THREE.SpriteMaterial({
|
||||
map: texture,
|
||||
transparent: true,
|
||||
depthTest: true,
|
||||
sizeAttenuation: true,
|
||||
});
|
||||
|
||||
console.log('[LODSystem] Initialized');
|
||||
}
|
||||
|
||||
function registerAvatar(avatarMesh, userId, color) {
|
||||
// Create billboard sprite for this avatar
|
||||
const spriteMat = _spriteMaterial.clone();
|
||||
if (color) {
|
||||
// Tint sprite to match avatar color
|
||||
const canvas = document.createElement('canvas');
|
||||
canvas.width = 64;
|
||||
canvas.height = 64;
|
||||
const ctx = canvas.getContext('2d');
|
||||
ctx.fillStyle = color;
|
||||
ctx.beginPath();
|
||||
ctx.arc(32, 32, 20, 0, Math.PI * 2);
|
||||
ctx.fill();
|
||||
ctx.fillStyle = '#0a0f1a';
|
||||
ctx.beginPath();
|
||||
ctx.arc(32, 28, 8, 0, Math.PI * 2);
|
||||
ctx.fill();
|
||||
spriteMat.map = new THREE.CanvasTexture(canvas);
|
||||
spriteMat.map.needsUpdate = true;
|
||||
}
|
||||
|
||||
const sprite = new THREE.Sprite(spriteMat);
|
||||
sprite.scale.set(SPRITE_SIZE, SPRITE_SIZE, 1);
|
||||
sprite.visible = false;
|
||||
_scene.add(sprite);
|
||||
|
||||
_registered.set(userId, {
|
||||
mesh: avatarMesh,
|
||||
sprite: sprite,
|
||||
distance: Infinity,
|
||||
});
|
||||
}
|
||||
|
||||
function unregisterAvatar(userId) {
|
||||
const entry = _registered.get(userId);
|
||||
if (entry) {
|
||||
_scene.remove(entry.sprite);
|
||||
entry.sprite.material.dispose();
|
||||
_registered.delete(userId);
|
||||
}
|
||||
}
|
||||
|
||||
function setSpriteColor(userId, color) {
|
||||
const entry = _registered.get(userId);
|
||||
if (!entry) return;
|
||||
const canvas = document.createElement('canvas');
|
||||
canvas.width = 64;
|
||||
canvas.height = 64;
|
||||
const ctx = canvas.getContext('2d');
|
||||
ctx.fillStyle = color;
|
||||
ctx.beginPath();
|
||||
ctx.arc(32, 32, 20, 0, Math.PI * 2);
|
||||
ctx.fill();
|
||||
ctx.fillStyle = '#0a0f1a';
|
||||
ctx.beginPath();
|
||||
ctx.arc(32, 28, 8, 0, Math.PI * 2);
|
||||
ctx.fill();
|
||||
entry.sprite.material.map = new THREE.CanvasTexture(canvas);
|
||||
entry.sprite.material.map.needsUpdate = true;
|
||||
}
|
||||
|
||||
function update(playerPos) {
|
||||
if (!_camera) return;
|
||||
|
||||
// Update frustum for culling
|
||||
_projScreenMatrix.multiplyMatrices(
|
||||
_camera.projectionMatrix,
|
||||
_camera.matrixWorldInverse
|
||||
);
|
||||
_frustum.setFromProjectionMatrix(_projScreenMatrix);
|
||||
|
||||
_registered.forEach((entry, userId) => {
|
||||
if (!entry.mesh) return;
|
||||
|
||||
const meshPos = entry.mesh.position;
|
||||
const distance = playerPos.distanceTo(meshPos);
|
||||
entry.distance = distance;
|
||||
|
||||
// Beyond cull distance: hide everything
|
||||
if (distance > LOD_CULL) {
|
||||
entry.mesh.visible = false;
|
||||
entry.sprite.visible = false;
|
||||
return;
|
||||
}
|
||||
|
||||
// Check if in camera frustum
|
||||
const inFrustum = _frustum.containsPoint(meshPos);
|
||||
if (!inFrustum) {
|
||||
entry.mesh.visible = false;
|
||||
entry.sprite.visible = false;
|
||||
return;
|
||||
}
|
||||
|
||||
// LOD switching
|
||||
if (distance <= LOD_NEAR) {
|
||||
// Near: full mesh
|
||||
entry.mesh.visible = true;
|
||||
entry.sprite.visible = false;
|
||||
} else if (distance <= LOD_FAR) {
|
||||
// Mid: mesh with reduced detail (keep mesh visible)
|
||||
entry.mesh.visible = true;
|
||||
entry.sprite.visible = false;
|
||||
} else {
|
||||
// Far: billboard sprite
|
||||
entry.mesh.visible = false;
|
||||
entry.sprite.visible = true;
|
||||
entry.sprite.position.copy(meshPos);
|
||||
entry.sprite.position.y += 1.2; // above avatar center
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
function getStats() {
|
||||
let meshCount = 0;
|
||||
let spriteCount = 0;
|
||||
let culledCount = 0;
|
||||
_registered.forEach(entry => {
|
||||
if (entry.mesh.visible) meshCount++;
|
||||
else if (entry.sprite.visible) spriteCount++;
|
||||
else culledCount++;
|
||||
});
|
||||
return { total: _registered.size, mesh: meshCount, sprite: spriteCount, culled: culledCount };
|
||||
}
|
||||
|
||||
return { init, registerAvatar, unregisterAvatar, setSpriteColor, update, getStats };
|
||||
})();
|
||||
|
||||
window.LODSystem = LODSystem;
|
||||
36
preview/nginx.conf
Normal file
36
preview/nginx.conf
Normal file
@@ -0,0 +1,36 @@
|
||||
server {
|
||||
listen 3000;
|
||||
server_name _;
|
||||
root /usr/share/nginx/html;
|
||||
index index.html;
|
||||
|
||||
location / {
|
||||
try_files $uri $uri/ /index.html;
|
||||
}
|
||||
|
||||
location ~* \.js$ {
|
||||
types { application/javascript js; }
|
||||
}
|
||||
|
||||
location ~* \.mjs$ {
|
||||
types { application/javascript mjs; }
|
||||
}
|
||||
|
||||
location ~* \.css$ {
|
||||
types { text/css css; }
|
||||
}
|
||||
|
||||
location ~* \.json$ {
|
||||
types { application/json json; }
|
||||
add_header Cache-Control "no-cache";
|
||||
}
|
||||
|
||||
location /api/world/ws {
|
||||
proxy_pass http://nexus-main:8765;
|
||||
proxy_http_version 1.1;
|
||||
proxy_set_header Upgrade $http_upgrade;
|
||||
proxy_set_header Connection "upgrade";
|
||||
proxy_set_header Host $host;
|
||||
proxy_read_timeout 86400;
|
||||
}
|
||||
}
|
||||
@@ -4,61 +4,48 @@ Sync branch protection rules from .gitea/branch-protection/*.yml to Gitea.
|
||||
Correctly uses the Gitea 1.25+ API (not GitHub-style).
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import json
|
||||
import os
|
||||
import sys
|
||||
import json
|
||||
import urllib.request
|
||||
from pathlib import Path
|
||||
|
||||
import yaml
|
||||
|
||||
GITEA_URL = os.getenv("GITEA_URL", "https://forge.alexanderwhitestone.com")
|
||||
GITEA_TOKEN = os.getenv("GITEA_TOKEN", "")
|
||||
ORG = "Timmy_Foundation"
|
||||
PROJECT_ROOT = Path(__file__).resolve().parent.parent
|
||||
CONFIG_DIR = PROJECT_ROOT / ".gitea" / "branch-protection"
|
||||
CONFIG_DIR = ".gitea/branch-protection"
|
||||
|
||||
|
||||
def api_request(method: str, path: str, payload: dict | None = None) -> dict:
|
||||
url = f"{GITEA_URL}/api/v1{path}"
|
||||
data = json.dumps(payload).encode() if payload else None
|
||||
req = urllib.request.Request(
|
||||
url,
|
||||
data=data,
|
||||
method=method,
|
||||
headers={
|
||||
"Authorization": f"token {GITEA_TOKEN}",
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
)
|
||||
req = urllib.request.Request(url, data=data, method=method, headers={
|
||||
"Authorization": f"token {GITEA_TOKEN}",
|
||||
"Content-Type": "application/json",
|
||||
})
|
||||
with urllib.request.urlopen(req, timeout=30) as resp:
|
||||
return json.loads(resp.read().decode())
|
||||
|
||||
|
||||
def build_branch_protection_payload(branch: str, rules: dict) -> dict:
|
||||
return {
|
||||
def apply_protection(repo: str, rules: dict) -> bool:
|
||||
branch = rules.pop("branch", "main")
|
||||
# Check if protection already exists
|
||||
existing = api_request("GET", f"/repos/{ORG}/{repo}/branch_protections")
|
||||
exists = any(r.get("branch_name") == branch for r in existing)
|
||||
|
||||
payload = {
|
||||
"branch_name": branch,
|
||||
"rule_name": branch,
|
||||
"required_approvals": rules.get("required_approvals", 1),
|
||||
"block_on_rejected_reviews": rules.get("block_on_rejected_reviews", True),
|
||||
"dismiss_stale_approvals": rules.get("dismiss_stale_approvals", True),
|
||||
"block_deletions": rules.get("block_deletions", True),
|
||||
"block_force_push": rules.get("block_force_push", rules.get("block_force_pushes", True)),
|
||||
"block_force_push": rules.get("block_force_push", True),
|
||||
"block_admin_merge_override": rules.get("block_admin_merge_override", True),
|
||||
"enable_status_check": rules.get("require_ci_to_merge", False),
|
||||
"status_check_contexts": rules.get("status_check_contexts", []),
|
||||
"block_on_outdated_branch": rules.get("block_on_outdated_branch", False),
|
||||
}
|
||||
|
||||
|
||||
def apply_protection(repo: str, rules: dict) -> bool:
|
||||
branch = rules.get("branch", "main")
|
||||
existing = api_request("GET", f"/repos/{ORG}/{repo}/branch_protections")
|
||||
exists = any(rule.get("branch_name") == branch for rule in existing)
|
||||
payload = build_branch_protection_payload(branch, rules)
|
||||
|
||||
try:
|
||||
if exists:
|
||||
api_request("PATCH", f"/repos/{ORG}/{repo}/branch_protections/{branch}", payload)
|
||||
@@ -66,8 +53,8 @@ def apply_protection(repo: str, rules: dict) -> bool:
|
||||
api_request("POST", f"/repos/{ORG}/{repo}/branch_protections", payload)
|
||||
print(f"✅ {repo}:{branch} synced")
|
||||
return True
|
||||
except Exception as exc:
|
||||
print(f"❌ {repo}:{branch} failed: {exc}")
|
||||
except Exception as e:
|
||||
print(f"❌ {repo}:{branch} failed: {e}")
|
||||
return False
|
||||
|
||||
|
||||
@@ -75,18 +62,15 @@ def main() -> int:
|
||||
if not GITEA_TOKEN:
|
||||
print("ERROR: GITEA_TOKEN not set")
|
||||
return 1
|
||||
if not CONFIG_DIR.exists():
|
||||
print(f"ERROR: config directory not found: {CONFIG_DIR}")
|
||||
return 1
|
||||
|
||||
ok = 0
|
||||
for cfg_path in sorted(CONFIG_DIR.glob("*.yml")):
|
||||
repo = cfg_path.stem
|
||||
with cfg_path.open() as fh:
|
||||
cfg = yaml.safe_load(fh) or {}
|
||||
rules = cfg.get("rules", {})
|
||||
rules.setdefault("branch", cfg.get("branch", "main"))
|
||||
if apply_protection(repo, rules):
|
||||
for fname in os.listdir(CONFIG_DIR):
|
||||
if not fname.endswith(".yml"):
|
||||
continue
|
||||
repo = fname[:-4]
|
||||
with open(os.path.join(CONFIG_DIR, fname)) as f:
|
||||
cfg = yaml.safe_load(f)
|
||||
if apply_protection(repo, cfg.get("rules", {})):
|
||||
ok += 1
|
||||
|
||||
print(f"\nSynced {ok} repo(s)")
|
||||
|
||||
46
tests/test_preview_deploy.py
Normal file
46
tests/test_preview_deploy.py
Normal file
@@ -0,0 +1,46 @@
|
||||
from pathlib import Path
|
||||
|
||||
|
||||
ROOT = Path(__file__).resolve().parent.parent
|
||||
DOCKERFILE = ROOT / "Dockerfile.preview"
|
||||
NGINX_CONF = ROOT / "preview" / "nginx.conf"
|
||||
DOC = ROOT / "docs" / "preview-deploy.md"
|
||||
COMPOSE = ROOT / "docker-compose.yml"
|
||||
|
||||
|
||||
def test_preview_deploy_files_exist():
|
||||
assert DOCKERFILE.exists(), "expected Dockerfile.preview for Nexus preview deployment"
|
||||
assert NGINX_CONF.exists(), "expected preview/nginx.conf for Nexus preview deployment"
|
||||
assert DOC.exists(), "expected docs/preview-deploy.md runbook"
|
||||
|
||||
|
||||
def test_preview_nginx_config_proxies_websocket_and_serves_modules():
|
||||
text = NGINX_CONF.read_text(encoding="utf-8")
|
||||
assert "listen 3000;" in text
|
||||
assert "location /api/world/ws" in text
|
||||
assert "proxy_pass http://nexus-main:8765;" in text
|
||||
assert "application/javascript js;" in text
|
||||
assert "try_files $uri $uri/ /index.html;" in text
|
||||
|
||||
|
||||
def test_compose_exposes_preview_service():
|
||||
text = COMPOSE.read_text(encoding="utf-8")
|
||||
assert "nexus-preview:" in text
|
||||
assert '"3000:3000"' in text
|
||||
assert "depends_on:" in text
|
||||
assert "nexus-main" in text
|
||||
|
||||
|
||||
def test_preview_runbook_documents_preview_url():
|
||||
text = DOC.read_text(encoding="utf-8")
|
||||
assert "http://localhost:3000" in text
|
||||
assert "docker compose up -d nexus-main nexus-preview" in text
|
||||
assert "/api/world/ws" in text
|
||||
|
||||
|
||||
def test_app_avoids_hardcoded_preview_breaking_urls():
|
||||
text = (ROOT / "app.js").read_text(encoding="utf-8")
|
||||
assert "ws://143.198.27.163:8765" not in text
|
||||
assert "http://localhost:8082/metrics" not in text
|
||||
assert "const metricsUrl = metricsOverride || `${window.location.protocol}//${window.location.host}/metrics`;" in text
|
||||
assert "const wsStatusUrl = `${protocol}//${window.location.host}/api/world/ws`;" in text
|
||||
@@ -1,45 +0,0 @@
|
||||
from __future__ import annotations
|
||||
|
||||
import importlib.util
|
||||
import sys
|
||||
from pathlib import Path
|
||||
|
||||
import yaml
|
||||
|
||||
PROJECT_ROOT = Path(__file__).parent.parent
|
||||
|
||||
_spec = importlib.util.spec_from_file_location(
|
||||
"sync_branch_protection_test",
|
||||
PROJECT_ROOT / "scripts" / "sync_branch_protection.py",
|
||||
)
|
||||
_mod = importlib.util.module_from_spec(_spec)
|
||||
sys.modules["sync_branch_protection_test"] = _mod
|
||||
_spec.loader.exec_module(_mod)
|
||||
|
||||
build_branch_protection_payload = _mod.build_branch_protection_payload
|
||||
|
||||
|
||||
def test_build_branch_protection_payload_enables_rebase_before_merge():
|
||||
payload = build_branch_protection_payload(
|
||||
"main",
|
||||
{
|
||||
"required_approvals": 1,
|
||||
"dismiss_stale_approvals": True,
|
||||
"require_ci_to_merge": False,
|
||||
"block_deletions": True,
|
||||
"block_force_push": True,
|
||||
"block_on_outdated_branch": True,
|
||||
},
|
||||
)
|
||||
|
||||
assert payload["branch_name"] == "main"
|
||||
assert payload["rule_name"] == "main"
|
||||
assert payload["block_on_outdated_branch"] is True
|
||||
assert payload["required_approvals"] == 1
|
||||
assert payload["enable_status_check"] is False
|
||||
|
||||
|
||||
def test_the_nexus_branch_protection_config_requires_up_to_date_branch():
|
||||
config = yaml.safe_load((PROJECT_ROOT / ".gitea" / "branch-protection" / "the-nexus.yml").read_text())
|
||||
rules = config["rules"]
|
||||
assert rules["block_on_outdated_branch"] is True
|
||||
@@ -1,236 +0,0 @@
|
||||
#!/usr/bin/env python3
|
||||
"""WebSocket Load Test — Measure concurrent connection capacity.
|
||||
|
||||
Simulates N concurrent WebSocket connections to the Nexus gateway
|
||||
and measures latency, throughput, and memory under load.
|
||||
|
||||
Usage:
|
||||
python3 tests/ws_load_test.py --url ws://localhost:8080 --connections 50
|
||||
python3 tests/ws_load_test.py --url ws://localhost:8080 --connections 100 --duration 30
|
||||
|
||||
Requirements: websockets (pip install websockets)
|
||||
"""
|
||||
|
||||
import argparse
|
||||
import asyncio
|
||||
import json
|
||||
import os
|
||||
import sys
|
||||
import time
|
||||
from dataclasses import dataclass, field
|
||||
from pathlib import Path
|
||||
|
||||
try:
|
||||
import websockets
|
||||
except ImportError:
|
||||
print("ERROR: websockets not installed. Run: pip install websockets")
|
||||
sys.exit(1)
|
||||
|
||||
|
||||
@dataclass
|
||||
class ConnectionStats:
|
||||
"""Stats for a single WebSocket connection."""
|
||||
connected: bool = False
|
||||
messages_sent: int = 0
|
||||
messages_received: int = 0
|
||||
errors: int = 0
|
||||
latencies: list = field(default_factory=list)
|
||||
connect_time: float = 0.0
|
||||
disconnect_time: float = 0.0
|
||||
|
||||
|
||||
@dataclass
|
||||
class LoadTestResults:
|
||||
"""Aggregate results for the load test."""
|
||||
total_connections: int = 0
|
||||
successful_connections: int = 0
|
||||
failed_connections: int = 0
|
||||
total_messages_sent: int = 0
|
||||
total_messages_received: int = 0
|
||||
total_errors: int = 0
|
||||
latencies: list = field(default_factory=list)
|
||||
duration: float = 0.0
|
||||
peak_memory_mb: float = 0.0
|
||||
|
||||
|
||||
async def connect_and_test(
|
||||
url: str,
|
||||
client_id: int,
|
||||
duration: int,
|
||||
message_interval: float,
|
||||
stats: ConnectionStats,
|
||||
results: LoadTestResults,
|
||||
):
|
||||
"""Single client: connect, send messages, measure responses."""
|
||||
start = time.time()
|
||||
try:
|
||||
async with websockets.connect(url, open_timeout=10) as ws:
|
||||
stats.connected = True
|
||||
stats.connect_time = time.time() - start
|
||||
results.successful_connections += 1
|
||||
|
||||
# Send a test message
|
||||
test_msg = json.dumps({
|
||||
"type": "ping",
|
||||
"client_id": client_id,
|
||||
"timestamp": time.time(),
|
||||
})
|
||||
|
||||
end_time = time.time() + duration
|
||||
while time.time() < end_time:
|
||||
try:
|
||||
send_time = time.time()
|
||||
await ws.send(test_msg)
|
||||
stats.messages_sent += 1
|
||||
results.total_messages_sent += 1
|
||||
|
||||
# Wait for response
|
||||
response = await asyncio.wait_for(ws.recv(), timeout=5.0)
|
||||
recv_time = time.time()
|
||||
latency = (recv_time - send_time) * 1000 # ms
|
||||
stats.latencies.append(latency)
|
||||
results.latencies.append(latency)
|
||||
stats.messages_received += 1
|
||||
results.total_messages_received += 1
|
||||
|
||||
except asyncio.TimeoutError:
|
||||
stats.errors += 1
|
||||
results.total_errors += 1
|
||||
except Exception as e:
|
||||
stats.errors += 1
|
||||
results.total_errors += 1
|
||||
|
||||
await asyncio.sleep(message_interval)
|
||||
|
||||
except Exception as e:
|
||||
stats.connected = False
|
||||
stats.errors += 1
|
||||
results.failed_connections += 1
|
||||
results.total_errors += 1
|
||||
|
||||
stats.disconnect_time = time.time()
|
||||
|
||||
|
||||
def get_memory_mb() -> float:
|
||||
"""Get current process memory in MB."""
|
||||
try:
|
||||
import psutil
|
||||
return psutil.Process().memory_info().rss / 1024 / 1024
|
||||
except ImportError:
|
||||
return 0.0
|
||||
|
||||
|
||||
async def run_load_test(
|
||||
url: str,
|
||||
num_connections: int,
|
||||
duration: int,
|
||||
message_interval: float,
|
||||
) -> LoadTestResults:
|
||||
"""Run the load test with N concurrent connections."""
|
||||
results = LoadTestResults(total_connections=num_connections)
|
||||
stats_list = [ConnectionStats() for _ in range(num_connections)]
|
||||
|
||||
print(f"Starting load test: {num_connections} connections to {url}")
|
||||
print(f"Duration: {duration}s, Message interval: {message_interval}s")
|
||||
print()
|
||||
|
||||
start_time = time.time()
|
||||
start_memory = get_memory_mb()
|
||||
|
||||
# Launch all connections concurrently
|
||||
tasks = [
|
||||
connect_and_test(
|
||||
url=url,
|
||||
client_id=i,
|
||||
duration=duration,
|
||||
message_interval=message_interval,
|
||||
stats=stats_list[i],
|
||||
results=results,
|
||||
)
|
||||
for i in range(num_connections)
|
||||
]
|
||||
|
||||
await asyncio.gather(*tasks, return_exceptions=True)
|
||||
|
||||
end_time = time.time()
|
||||
end_memory = get_memory_mb()
|
||||
|
||||
results.duration = end_time - start_time
|
||||
results.peak_memory_mb = max(start_memory, end_memory)
|
||||
|
||||
return results
|
||||
|
||||
|
||||
def print_results(results: LoadTestResults):
|
||||
"""Print load test results."""
|
||||
print()
|
||||
print("=" * 60)
|
||||
print("WEBSOCKET LOAD TEST RESULTS")
|
||||
print("=" * 60)
|
||||
print(f"Connections: {results.total_connections}")
|
||||
print(f"Successful: {results.successful_connections}")
|
||||
print(f"Failed: {results.failed_connections}")
|
||||
print(f"Duration: {results.duration:.1f}s")
|
||||
print()
|
||||
print(f"Messages sent: {results.total_messages_sent}")
|
||||
print(f"Messages recv: {results.total_messages_received}")
|
||||
print(f"Errors: {results.total_errors}")
|
||||
print(f"Throughput: {results.total_messages_sent / max(results.duration, 1):.1f} msg/s")
|
||||
print()
|
||||
|
||||
if results.latencies:
|
||||
results.latencies.sort()
|
||||
n = len(results.latencies)
|
||||
print(f"Latency (ms):")
|
||||
print(f" p50: {results.latencies[n // 2]:.1f}")
|
||||
print(f" p90: {results.latencies[int(n * 0.9)]:.1f}")
|
||||
print(f" p95: {results.latencies[int(n * 0.95)]:.1f}")
|
||||
print(f" p99: {results.latencies[min(int(n * 0.99), n-1)]:.1f}")
|
||||
print(f" max: {results.latencies[-1]:.1f}")
|
||||
print(f" mean: {sum(results.latencies) / n:.1f}")
|
||||
|
||||
print()
|
||||
print(f"Memory delta: {results.peak_memory_mb:.1f} MB")
|
||||
print("=" * 60)
|
||||
|
||||
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(description="WebSocket load test")
|
||||
parser.add_argument("--url", default="ws://localhost:8080", help="WebSocket URL")
|
||||
parser.add_argument("--connections", type=int, default=10, help="Number of concurrent connections")
|
||||
parser.add_argument("--duration", type=int, default=10, help="Test duration in seconds")
|
||||
parser.add_argument("--interval", type=float, default=0.5, help="Message interval in seconds")
|
||||
parser.add_argument("--output", help="Save results to JSON file")
|
||||
args = parser.parse_args()
|
||||
|
||||
results = asyncio.run(run_load_test(
|
||||
url=args.url,
|
||||
num_connections=args.connections,
|
||||
duration=args.duration,
|
||||
message_interval=args.interval,
|
||||
))
|
||||
|
||||
print_results(results)
|
||||
|
||||
if args.output:
|
||||
data = {
|
||||
"url": args.url,
|
||||
"connections": args.connections,
|
||||
"duration": args.duration,
|
||||
"interval": args.interval,
|
||||
"total_connections": results.total_connections,
|
||||
"successful": results.successful_connections,
|
||||
"failed": results.failed_connections,
|
||||
"messages_sent": results.total_messages_sent,
|
||||
"messages_received": results.total_messages_received,
|
||||
"errors": results.total_errors,
|
||||
"duration_seconds": results.duration,
|
||||
"memory_mb": results.peak_memory_mb,
|
||||
}
|
||||
with open(args.output, "w") as f:
|
||||
json.dump(data, f, indent=2)
|
||||
print(f"\nResults saved to {args.output}")
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
Reference in New Issue
Block a user