Compare commits

..

4 Commits

Author SHA1 Message Date
Alexander Whitestone
ee2be0427c feat: add Nexus preview URL deployment stack (#1339)
Some checks failed
CI / test (pull_request) Failing after 59s
CI / validate (pull_request) Failing after 34s
Review Approval Gate / verify-review (pull_request) Successful in 6s
2026-04-15 03:40:44 -04:00
Alexander Whitestone
5fb8c0c513 wip: wire preview service and dynamic preview URLs 2026-04-15 03:39:58 -04:00
Alexander Whitestone
a796453766 wip: add preview deploy stack artifacts 2026-04-15 03:39:15 -04:00
Alexander Whitestone
b4b029d2a6 wip: add preview deploy regression test 2026-04-15 03:35:20 -04:00
13 changed files with 161 additions and 524 deletions

View File

@@ -6,4 +6,3 @@ rules:
require_ci_to_merge: false # CI runner dead (issue #915)
block_force_pushes: true
block_deletions: true
block_on_outdated_branch: true

View File

@@ -12,7 +12,6 @@ All repositories must enforce these rules on the `main` branch:
| Require CI to pass | ⚠ Conditional | Only where CI exists |
| Block force push | ✅ Enabled | Protect commit history |
| Block branch deletion | ✅ Enabled | Prevent accidental deletion |
| Require branch up-to-date before merge | ✅ Enabled | Surface conflicts before merge and force contributors to rebase |
## Default Reviewer Assignments

9
Dockerfile.preview Normal file
View File

@@ -0,0 +1,9 @@
FROM nginx:alpine
RUN rm /etc/nginx/conf.d/default.conf
COPY preview/nginx.conf /etc/nginx/conf.d/default.conf
COPY *.html *.js *.mjs *.json *.css /usr/share/nginx/html/
COPY nexus/ /usr/share/nginx/html/nexus/
EXPOSE 3000

20
app.js
View File

@@ -714,10 +714,6 @@ async function init() {
camera = new THREE.PerspectiveCamera(65, window.innerWidth / window.innerHeight, 0.1, 1000);
camera.position.copy(playerPos);
// Initialize avatar and LOD systems
if (window.AvatarCustomization) window.AvatarCustomization.init(scene, camera);
if (window.LODSystem) window.LODSystem.init(scene, camera);
updateLoad(20);
createSkybox();
@@ -1253,10 +1249,16 @@ async function updateSovereignHealth() {
const container = document.getElementById('sovereign-health-content');
if (!container) return;
const params = new URLSearchParams(window.location.search);
const metricsOverride = params.get('metrics');
const metricsUrl = metricsOverride || `${window.location.protocol}//${window.location.host}/metrics`;
const protocol = window.location.protocol === 'https:' ? 'wss:' : 'ws:';
const wsStatusUrl = `${protocol}//${window.location.host}/api/world/ws`;
let metrics = { sovereignty_score: 100, local_sessions: 0, total_sessions: 0 };
let daemonReachable = false;
try {
const res = await fetch('http://localhost:8082/metrics');
const res = await fetch(metricsUrl);
if (res.ok) {
metrics = await res.json();
daemonReachable = true;
@@ -1269,8 +1271,8 @@ async function updateSovereignHealth() {
{ name: 'LOCAL DAEMON', status: daemonReachable ? 'ONLINE' : 'OFFLINE' },
{ name: 'FORGE / GITEA', url: 'https://forge.alexanderwhitestone.com', status: 'ONLINE' },
{ name: 'NEXUS CORE', url: 'https://forge.alexanderwhitestone.com/Timmy_Foundation/the-nexus', status: 'ONLINE' },
{ name: 'HERMES WS', url: 'ws://143.198.27.163:8765', status: wsConnected ? 'ONLINE' : 'OFFLINE' },
{ name: 'SOVEREIGNTY', url: 'http://localhost:8082/metrics', status: metrics.sovereignty_score + '%' }
{ name: 'HERMES WS', url: wsStatusUrl, status: wsConnected ? 'ONLINE' : 'OFFLINE' },
{ name: 'SOVEREIGNTY', url: metricsUrl, status: metrics.sovereignty_score + '%' }
];
container.innerHTML = '';
@@ -3561,10 +3563,6 @@ function gameLoop() {
if (composer) { composer.render(); } else { renderer.render(scene, camera); }
// Update avatar and LOD systems
if (window.AvatarCustomization && playerPos) window.AvatarCustomization.update(playerPos);
if (window.LODSystem && playerPos) window.LODSystem.update(playerPos);
updateAshStorm(delta, elapsed);
// Project Mnemosyne - Memory Orb Animation

View File

@@ -1,5 +1,3 @@
version: "3.9"
services:
nexus-main:
build: .
@@ -7,9 +5,21 @@ services:
restart: unless-stopped
ports:
- "8765:8765"
nexus-staging:
build: .
container_name: nexus-staging
restart: unless-stopped
ports:
- "8766:8765"
- "8766:8765"
nexus-preview:
build:
context: .
dockerfile: Dockerfile.preview
container_name: nexus-preview
restart: unless-stopped
ports:
- "3000:3000"
depends_on:
- nexus-main

25
docs/preview-deploy.md Normal file
View File

@@ -0,0 +1,25 @@
# Nexus preview deploy
The Nexus frontend must be served over HTTP for ES modules to boot. This repo now includes a preview stack that serves the frontend on a proper URL and proxies `/api/world/ws` back to the existing Nexus WebSocket gateway.
## Quick start
```bash
docker compose up -d nexus-main nexus-preview
```
Open:
- `http://localhost:3000`
The preview service serves the static frontend and proxies WebSocket traffic at:
- `/api/world/ws`
## Remote preview
If you run the same compose stack on a VPS, the preview URL is:
- `http://<host>:3000`
## Notes
- `nexus-main` keeps serving the backend WebSocket gateway on port `8765`
- `nexus-preview` serves the frontend on port `3000`
- The browser can stay on a single origin because nginx proxies the WebSocket path

View File

@@ -395,8 +395,6 @@
<div id="memory-connections-panel" class="memory-connections-panel" style="display:none;" aria-label="Memory Connections Panel"></div>
<script src="./boot.js"></script>
<script src="./avatar-customization.js"></script>
<script src="./lod-system.js"></script>
<script>
function openMemoryFilter() { renderFilterList(); document.getElementById('memory-filter').style.display = 'flex'; }
function closeMemoryFilter() { document.getElementById('memory-filter').style.display = 'none'; }

View File

@@ -1,186 +0,0 @@
/**
* LOD (Level of Detail) System for The Nexus
*
* Optimizes rendering when many avatars/users are visible:
* - Distance-based LOD: far users become billboard sprites
* - Occlusion: skip rendering users behind walls
* - Budget: maintain 60 FPS target with 50+ avatars
*
* Usage:
* LODSystem.init(scene, camera);
* LODSystem.registerAvatar(avatarMesh, userId);
* LODSystem.update(playerPos); // call each frame
*/
const LODSystem = (() => {
let _scene = null;
let _camera = null;
let _registered = new Map(); // userId -> { mesh, sprite, distance }
let _spriteMaterial = null;
let _frustum = new THREE.Frustum();
let _projScreenMatrix = new THREE.Matrix4();
// Thresholds
const LOD_NEAR = 15; // Full mesh within 15 units
const LOD_FAR = 40; // Billboard beyond 40 units
const LOD_CULL = 80; // Don't render beyond 80 units
const SPRITE_SIZE = 1.2;
function init(sceneRef, cameraRef) {
_scene = sceneRef;
_camera = cameraRef;
// Create shared sprite material
const canvas = document.createElement('canvas');
canvas.width = 64;
canvas.height = 64;
const ctx = canvas.getContext('2d');
// Simple avatar indicator: colored circle
ctx.fillStyle = '#00ffcc';
ctx.beginPath();
ctx.arc(32, 32, 20, 0, Math.PI * 2);
ctx.fill();
ctx.fillStyle = '#0a0f1a';
ctx.beginPath();
ctx.arc(32, 28, 8, 0, Math.PI * 2); // head
ctx.fill();
const texture = new THREE.CanvasTexture(canvas);
_spriteMaterial = new THREE.SpriteMaterial({
map: texture,
transparent: true,
depthTest: true,
sizeAttenuation: true,
});
console.log('[LODSystem] Initialized');
}
function registerAvatar(avatarMesh, userId, color) {
// Create billboard sprite for this avatar
const spriteMat = _spriteMaterial.clone();
if (color) {
// Tint sprite to match avatar color
const canvas = document.createElement('canvas');
canvas.width = 64;
canvas.height = 64;
const ctx = canvas.getContext('2d');
ctx.fillStyle = color;
ctx.beginPath();
ctx.arc(32, 32, 20, 0, Math.PI * 2);
ctx.fill();
ctx.fillStyle = '#0a0f1a';
ctx.beginPath();
ctx.arc(32, 28, 8, 0, Math.PI * 2);
ctx.fill();
spriteMat.map = new THREE.CanvasTexture(canvas);
spriteMat.map.needsUpdate = true;
}
const sprite = new THREE.Sprite(spriteMat);
sprite.scale.set(SPRITE_SIZE, SPRITE_SIZE, 1);
sprite.visible = false;
_scene.add(sprite);
_registered.set(userId, {
mesh: avatarMesh,
sprite: sprite,
distance: Infinity,
});
}
function unregisterAvatar(userId) {
const entry = _registered.get(userId);
if (entry) {
_scene.remove(entry.sprite);
entry.sprite.material.dispose();
_registered.delete(userId);
}
}
function setSpriteColor(userId, color) {
const entry = _registered.get(userId);
if (!entry) return;
const canvas = document.createElement('canvas');
canvas.width = 64;
canvas.height = 64;
const ctx = canvas.getContext('2d');
ctx.fillStyle = color;
ctx.beginPath();
ctx.arc(32, 32, 20, 0, Math.PI * 2);
ctx.fill();
ctx.fillStyle = '#0a0f1a';
ctx.beginPath();
ctx.arc(32, 28, 8, 0, Math.PI * 2);
ctx.fill();
entry.sprite.material.map = new THREE.CanvasTexture(canvas);
entry.sprite.material.map.needsUpdate = true;
}
function update(playerPos) {
if (!_camera) return;
// Update frustum for culling
_projScreenMatrix.multiplyMatrices(
_camera.projectionMatrix,
_camera.matrixWorldInverse
);
_frustum.setFromProjectionMatrix(_projScreenMatrix);
_registered.forEach((entry, userId) => {
if (!entry.mesh) return;
const meshPos = entry.mesh.position;
const distance = playerPos.distanceTo(meshPos);
entry.distance = distance;
// Beyond cull distance: hide everything
if (distance > LOD_CULL) {
entry.mesh.visible = false;
entry.sprite.visible = false;
return;
}
// Check if in camera frustum
const inFrustum = _frustum.containsPoint(meshPos);
if (!inFrustum) {
entry.mesh.visible = false;
entry.sprite.visible = false;
return;
}
// LOD switching
if (distance <= LOD_NEAR) {
// Near: full mesh
entry.mesh.visible = true;
entry.sprite.visible = false;
} else if (distance <= LOD_FAR) {
// Mid: mesh with reduced detail (keep mesh visible)
entry.mesh.visible = true;
entry.sprite.visible = false;
} else {
// Far: billboard sprite
entry.mesh.visible = false;
entry.sprite.visible = true;
entry.sprite.position.copy(meshPos);
entry.sprite.position.y += 1.2; // above avatar center
}
});
}
function getStats() {
let meshCount = 0;
let spriteCount = 0;
let culledCount = 0;
_registered.forEach(entry => {
if (entry.mesh.visible) meshCount++;
else if (entry.sprite.visible) spriteCount++;
else culledCount++;
});
return { total: _registered.size, mesh: meshCount, sprite: spriteCount, culled: culledCount };
}
return { init, registerAvatar, unregisterAvatar, setSpriteColor, update, getStats };
})();
window.LODSystem = LODSystem;

36
preview/nginx.conf Normal file
View File

@@ -0,0 +1,36 @@
server {
listen 3000;
server_name _;
root /usr/share/nginx/html;
index index.html;
location / {
try_files $uri $uri/ /index.html;
}
location ~* \.js$ {
types { application/javascript js; }
}
location ~* \.mjs$ {
types { application/javascript mjs; }
}
location ~* \.css$ {
types { text/css css; }
}
location ~* \.json$ {
types { application/json json; }
add_header Cache-Control "no-cache";
}
location /api/world/ws {
proxy_pass http://nexus-main:8765;
proxy_http_version 1.1;
proxy_set_header Upgrade $http_upgrade;
proxy_set_header Connection "upgrade";
proxy_set_header Host $host;
proxy_read_timeout 86400;
}
}

View File

@@ -4,61 +4,48 @@ Sync branch protection rules from .gitea/branch-protection/*.yml to Gitea.
Correctly uses the Gitea 1.25+ API (not GitHub-style).
"""
from __future__ import annotations
import json
import os
import sys
import json
import urllib.request
from pathlib import Path
import yaml
GITEA_URL = os.getenv("GITEA_URL", "https://forge.alexanderwhitestone.com")
GITEA_TOKEN = os.getenv("GITEA_TOKEN", "")
ORG = "Timmy_Foundation"
PROJECT_ROOT = Path(__file__).resolve().parent.parent
CONFIG_DIR = PROJECT_ROOT / ".gitea" / "branch-protection"
CONFIG_DIR = ".gitea/branch-protection"
def api_request(method: str, path: str, payload: dict | None = None) -> dict:
url = f"{GITEA_URL}/api/v1{path}"
data = json.dumps(payload).encode() if payload else None
req = urllib.request.Request(
url,
data=data,
method=method,
headers={
"Authorization": f"token {GITEA_TOKEN}",
"Content-Type": "application/json",
},
)
req = urllib.request.Request(url, data=data, method=method, headers={
"Authorization": f"token {GITEA_TOKEN}",
"Content-Type": "application/json",
})
with urllib.request.urlopen(req, timeout=30) as resp:
return json.loads(resp.read().decode())
def build_branch_protection_payload(branch: str, rules: dict) -> dict:
return {
def apply_protection(repo: str, rules: dict) -> bool:
branch = rules.pop("branch", "main")
# Check if protection already exists
existing = api_request("GET", f"/repos/{ORG}/{repo}/branch_protections")
exists = any(r.get("branch_name") == branch for r in existing)
payload = {
"branch_name": branch,
"rule_name": branch,
"required_approvals": rules.get("required_approvals", 1),
"block_on_rejected_reviews": rules.get("block_on_rejected_reviews", True),
"dismiss_stale_approvals": rules.get("dismiss_stale_approvals", True),
"block_deletions": rules.get("block_deletions", True),
"block_force_push": rules.get("block_force_push", rules.get("block_force_pushes", True)),
"block_force_push": rules.get("block_force_push", True),
"block_admin_merge_override": rules.get("block_admin_merge_override", True),
"enable_status_check": rules.get("require_ci_to_merge", False),
"status_check_contexts": rules.get("status_check_contexts", []),
"block_on_outdated_branch": rules.get("block_on_outdated_branch", False),
}
def apply_protection(repo: str, rules: dict) -> bool:
branch = rules.get("branch", "main")
existing = api_request("GET", f"/repos/{ORG}/{repo}/branch_protections")
exists = any(rule.get("branch_name") == branch for rule in existing)
payload = build_branch_protection_payload(branch, rules)
try:
if exists:
api_request("PATCH", f"/repos/{ORG}/{repo}/branch_protections/{branch}", payload)
@@ -66,8 +53,8 @@ def apply_protection(repo: str, rules: dict) -> bool:
api_request("POST", f"/repos/{ORG}/{repo}/branch_protections", payload)
print(f"{repo}:{branch} synced")
return True
except Exception as exc:
print(f"{repo}:{branch} failed: {exc}")
except Exception as e:
print(f"{repo}:{branch} failed: {e}")
return False
@@ -75,18 +62,15 @@ def main() -> int:
if not GITEA_TOKEN:
print("ERROR: GITEA_TOKEN not set")
return 1
if not CONFIG_DIR.exists():
print(f"ERROR: config directory not found: {CONFIG_DIR}")
return 1
ok = 0
for cfg_path in sorted(CONFIG_DIR.glob("*.yml")):
repo = cfg_path.stem
with cfg_path.open() as fh:
cfg = yaml.safe_load(fh) or {}
rules = cfg.get("rules", {})
rules.setdefault("branch", cfg.get("branch", "main"))
if apply_protection(repo, rules):
for fname in os.listdir(CONFIG_DIR):
if not fname.endswith(".yml"):
continue
repo = fname[:-4]
with open(os.path.join(CONFIG_DIR, fname)) as f:
cfg = yaml.safe_load(f)
if apply_protection(repo, cfg.get("rules", {})):
ok += 1
print(f"\nSynced {ok} repo(s)")

View File

@@ -0,0 +1,46 @@
from pathlib import Path
ROOT = Path(__file__).resolve().parent.parent
DOCKERFILE = ROOT / "Dockerfile.preview"
NGINX_CONF = ROOT / "preview" / "nginx.conf"
DOC = ROOT / "docs" / "preview-deploy.md"
COMPOSE = ROOT / "docker-compose.yml"
def test_preview_deploy_files_exist():
assert DOCKERFILE.exists(), "expected Dockerfile.preview for Nexus preview deployment"
assert NGINX_CONF.exists(), "expected preview/nginx.conf for Nexus preview deployment"
assert DOC.exists(), "expected docs/preview-deploy.md runbook"
def test_preview_nginx_config_proxies_websocket_and_serves_modules():
text = NGINX_CONF.read_text(encoding="utf-8")
assert "listen 3000;" in text
assert "location /api/world/ws" in text
assert "proxy_pass http://nexus-main:8765;" in text
assert "application/javascript js;" in text
assert "try_files $uri $uri/ /index.html;" in text
def test_compose_exposes_preview_service():
text = COMPOSE.read_text(encoding="utf-8")
assert "nexus-preview:" in text
assert '"3000:3000"' in text
assert "depends_on:" in text
assert "nexus-main" in text
def test_preview_runbook_documents_preview_url():
text = DOC.read_text(encoding="utf-8")
assert "http://localhost:3000" in text
assert "docker compose up -d nexus-main nexus-preview" in text
assert "/api/world/ws" in text
def test_app_avoids_hardcoded_preview_breaking_urls():
text = (ROOT / "app.js").read_text(encoding="utf-8")
assert "ws://143.198.27.163:8765" not in text
assert "http://localhost:8082/metrics" not in text
assert "const metricsUrl = metricsOverride || `${window.location.protocol}//${window.location.host}/metrics`;" in text
assert "const wsStatusUrl = `${protocol}//${window.location.host}/api/world/ws`;" in text

View File

@@ -1,45 +0,0 @@
from __future__ import annotations
import importlib.util
import sys
from pathlib import Path
import yaml
PROJECT_ROOT = Path(__file__).parent.parent
_spec = importlib.util.spec_from_file_location(
"sync_branch_protection_test",
PROJECT_ROOT / "scripts" / "sync_branch_protection.py",
)
_mod = importlib.util.module_from_spec(_spec)
sys.modules["sync_branch_protection_test"] = _mod
_spec.loader.exec_module(_mod)
build_branch_protection_payload = _mod.build_branch_protection_payload
def test_build_branch_protection_payload_enables_rebase_before_merge():
payload = build_branch_protection_payload(
"main",
{
"required_approvals": 1,
"dismiss_stale_approvals": True,
"require_ci_to_merge": False,
"block_deletions": True,
"block_force_push": True,
"block_on_outdated_branch": True,
},
)
assert payload["branch_name"] == "main"
assert payload["rule_name"] == "main"
assert payload["block_on_outdated_branch"] is True
assert payload["required_approvals"] == 1
assert payload["enable_status_check"] is False
def test_the_nexus_branch_protection_config_requires_up_to_date_branch():
config = yaml.safe_load((PROJECT_ROOT / ".gitea" / "branch-protection" / "the-nexus.yml").read_text())
rules = config["rules"]
assert rules["block_on_outdated_branch"] is True

View File

@@ -1,236 +0,0 @@
#!/usr/bin/env python3
"""WebSocket Load Test — Measure concurrent connection capacity.
Simulates N concurrent WebSocket connections to the Nexus gateway
and measures latency, throughput, and memory under load.
Usage:
python3 tests/ws_load_test.py --url ws://localhost:8080 --connections 50
python3 tests/ws_load_test.py --url ws://localhost:8080 --connections 100 --duration 30
Requirements: websockets (pip install websockets)
"""
import argparse
import asyncio
import json
import os
import sys
import time
from dataclasses import dataclass, field
from pathlib import Path
try:
import websockets
except ImportError:
print("ERROR: websockets not installed. Run: pip install websockets")
sys.exit(1)
@dataclass
class ConnectionStats:
"""Stats for a single WebSocket connection."""
connected: bool = False
messages_sent: int = 0
messages_received: int = 0
errors: int = 0
latencies: list = field(default_factory=list)
connect_time: float = 0.0
disconnect_time: float = 0.0
@dataclass
class LoadTestResults:
"""Aggregate results for the load test."""
total_connections: int = 0
successful_connections: int = 0
failed_connections: int = 0
total_messages_sent: int = 0
total_messages_received: int = 0
total_errors: int = 0
latencies: list = field(default_factory=list)
duration: float = 0.0
peak_memory_mb: float = 0.0
async def connect_and_test(
url: str,
client_id: int,
duration: int,
message_interval: float,
stats: ConnectionStats,
results: LoadTestResults,
):
"""Single client: connect, send messages, measure responses."""
start = time.time()
try:
async with websockets.connect(url, open_timeout=10) as ws:
stats.connected = True
stats.connect_time = time.time() - start
results.successful_connections += 1
# Send a test message
test_msg = json.dumps({
"type": "ping",
"client_id": client_id,
"timestamp": time.time(),
})
end_time = time.time() + duration
while time.time() < end_time:
try:
send_time = time.time()
await ws.send(test_msg)
stats.messages_sent += 1
results.total_messages_sent += 1
# Wait for response
response = await asyncio.wait_for(ws.recv(), timeout=5.0)
recv_time = time.time()
latency = (recv_time - send_time) * 1000 # ms
stats.latencies.append(latency)
results.latencies.append(latency)
stats.messages_received += 1
results.total_messages_received += 1
except asyncio.TimeoutError:
stats.errors += 1
results.total_errors += 1
except Exception as e:
stats.errors += 1
results.total_errors += 1
await asyncio.sleep(message_interval)
except Exception as e:
stats.connected = False
stats.errors += 1
results.failed_connections += 1
results.total_errors += 1
stats.disconnect_time = time.time()
def get_memory_mb() -> float:
"""Get current process memory in MB."""
try:
import psutil
return psutil.Process().memory_info().rss / 1024 / 1024
except ImportError:
return 0.0
async def run_load_test(
url: str,
num_connections: int,
duration: int,
message_interval: float,
) -> LoadTestResults:
"""Run the load test with N concurrent connections."""
results = LoadTestResults(total_connections=num_connections)
stats_list = [ConnectionStats() for _ in range(num_connections)]
print(f"Starting load test: {num_connections} connections to {url}")
print(f"Duration: {duration}s, Message interval: {message_interval}s")
print()
start_time = time.time()
start_memory = get_memory_mb()
# Launch all connections concurrently
tasks = [
connect_and_test(
url=url,
client_id=i,
duration=duration,
message_interval=message_interval,
stats=stats_list[i],
results=results,
)
for i in range(num_connections)
]
await asyncio.gather(*tasks, return_exceptions=True)
end_time = time.time()
end_memory = get_memory_mb()
results.duration = end_time - start_time
results.peak_memory_mb = max(start_memory, end_memory)
return results
def print_results(results: LoadTestResults):
"""Print load test results."""
print()
print("=" * 60)
print("WEBSOCKET LOAD TEST RESULTS")
print("=" * 60)
print(f"Connections: {results.total_connections}")
print(f"Successful: {results.successful_connections}")
print(f"Failed: {results.failed_connections}")
print(f"Duration: {results.duration:.1f}s")
print()
print(f"Messages sent: {results.total_messages_sent}")
print(f"Messages recv: {results.total_messages_received}")
print(f"Errors: {results.total_errors}")
print(f"Throughput: {results.total_messages_sent / max(results.duration, 1):.1f} msg/s")
print()
if results.latencies:
results.latencies.sort()
n = len(results.latencies)
print(f"Latency (ms):")
print(f" p50: {results.latencies[n // 2]:.1f}")
print(f" p90: {results.latencies[int(n * 0.9)]:.1f}")
print(f" p95: {results.latencies[int(n * 0.95)]:.1f}")
print(f" p99: {results.latencies[min(int(n * 0.99), n-1)]:.1f}")
print(f" max: {results.latencies[-1]:.1f}")
print(f" mean: {sum(results.latencies) / n:.1f}")
print()
print(f"Memory delta: {results.peak_memory_mb:.1f} MB")
print("=" * 60)
def main():
parser = argparse.ArgumentParser(description="WebSocket load test")
parser.add_argument("--url", default="ws://localhost:8080", help="WebSocket URL")
parser.add_argument("--connections", type=int, default=10, help="Number of concurrent connections")
parser.add_argument("--duration", type=int, default=10, help="Test duration in seconds")
parser.add_argument("--interval", type=float, default=0.5, help="Message interval in seconds")
parser.add_argument("--output", help="Save results to JSON file")
args = parser.parse_args()
results = asyncio.run(run_load_test(
url=args.url,
num_connections=args.connections,
duration=args.duration,
message_interval=args.interval,
))
print_results(results)
if args.output:
data = {
"url": args.url,
"connections": args.connections,
"duration": args.duration,
"interval": args.interval,
"total_connections": results.total_connections,
"successful": results.successful_connections,
"failed": results.failed_connections,
"messages_sent": results.total_messages_sent,
"messages_received": results.total_messages_received,
"errors": results.total_errors,
"duration_seconds": results.duration,
"memory_mb": results.peak_memory_mb,
}
with open(args.output, "w") as f:
json.dump(data, f, indent=2)
print(f"\nResults saved to {args.output}")
if __name__ == "__main__":
main()