Compare commits
1 Commits
mimo/code/
...
mimo/build
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
c34748704e |
51
.gitea.yml
51
.gitea.yml
@@ -15,3 +15,54 @@ protection:
|
||||
- perplexity
|
||||
required_reviewers:
|
||||
- Timmy # Owner gate for hermes-agent
|
||||
main:
|
||||
require_pull_request: true
|
||||
required_approvals: 1
|
||||
dismiss_stale_approvals: true
|
||||
require_ci_to_pass: true
|
||||
block_force_push: true
|
||||
block_deletion: true
|
||||
>>>>>>> replace
|
||||
</source>
|
||||
|
||||
CODEOWNERS
|
||||
<source>
|
||||
<<<<<<< search
|
||||
protection:
|
||||
main:
|
||||
required_status_checks:
|
||||
- "ci/unit-tests"
|
||||
- "ci/integration"
|
||||
required_pull_request_reviews:
|
||||
- "1 approval"
|
||||
restrictions:
|
||||
- "block force push"
|
||||
- "block deletion"
|
||||
enforce_admins: true
|
||||
|
||||
the-nexus:
|
||||
required_status_checks: []
|
||||
required_pull_request_reviews:
|
||||
- "1 approval"
|
||||
restrictions:
|
||||
- "block force push"
|
||||
- "block deletion"
|
||||
enforce_admins: true
|
||||
|
||||
timmy-home:
|
||||
required_status_checks: []
|
||||
required_pull_request_reviews:
|
||||
- "1 approval"
|
||||
restrictions:
|
||||
- "block force push"
|
||||
- "block deletion"
|
||||
enforce_admins: true
|
||||
|
||||
timmy-config:
|
||||
required_status_checks: []
|
||||
required_pull_request_reviews:
|
||||
- "1 approval"
|
||||
restrictions:
|
||||
- "block force push"
|
||||
- "block deletion"
|
||||
enforce_admins: true
|
||||
|
||||
42
app.js
42
app.js
@@ -1,10 +1,9 @@
|
||||
import ResonanceVisualizer from './nexus/components/resonance-visualizer.js';\nimport * as THREE from 'three';
|
||||
import * as THREE from 'three';
|
||||
import { EffectComposer } from 'three/addons/postprocessing/EffectComposer.js';
|
||||
import { RenderPass } from 'three/addons/postprocessing/RenderPass.js';
|
||||
import { UnrealBloomPass } from 'three/addons/postprocessing/UnrealBloomPass.js';
|
||||
import { SMAAPass } from 'three/addons/postprocessing/SMAAPass.js';
|
||||
import { SpatialMemory } from './nexus/components/spatial-memory.js';
|
||||
import { SpatialAudio } from './nexus/components/spatial-audio.js';
|
||||
import { MemoryBirth } from './nexus/components/memory-birth.js';
|
||||
import { MemoryOptimizer } from './nexus/components/memory-optimizer.js';
|
||||
import { MemoryInspect } from './nexus/components/memory-inspect.js';
|
||||
@@ -598,7 +597,7 @@ class PSELayer {
|
||||
|
||||
let pseLayer;
|
||||
|
||||
let resonanceViz, metaLayer, neuroBridge, cbr, symbolicPlanner, knowledgeGraph, blackboard, symbolicEngine, calibrator;
|
||||
let metaLayer, neuroBridge, cbr, symbolicPlanner, knowledgeGraph, blackboard, symbolicEngine, calibrator;
|
||||
let agentFSMs = {};
|
||||
|
||||
function setupGOFAI() {
|
||||
@@ -667,7 +666,7 @@ async function init() {
|
||||
scene = new THREE.Scene();
|
||||
scene.fog = new THREE.FogExp2(0x050510, 0.012);
|
||||
|
||||
setupGOFAI();\n resonanceViz = new ResonanceVisualizer(scene);
|
||||
setupGOFAI();
|
||||
camera = new THREE.PerspectiveCamera(65, window.innerWidth / window.innerHeight, 0.1, 1000);
|
||||
camera.position.copy(playerPos);
|
||||
|
||||
@@ -705,19 +704,17 @@ async function init() {
|
||||
createParticles();
|
||||
createDustParticles();
|
||||
updateLoad(85);
|
||||
createAmbientStructures();
|
||||
if (performanceTier !== "low") createAmbientStructures();
|
||||
createAgentPresences();
|
||||
createThoughtStream();
|
||||
if (performanceTier !== "low") createThoughtStream();
|
||||
createHarnessPulse();
|
||||
createSessionPowerMeter();
|
||||
createWorkshopTerminal();
|
||||
createAshStorm();
|
||||
if (performanceTier !== "low") createAshStorm();
|
||||
SpatialMemory.init(scene);
|
||||
MemoryBirth.init(scene);
|
||||
MemoryBirth.wrapSpatialMemory(SpatialMemory);
|
||||
SpatialMemory.setCamera(camera);
|
||||
SpatialAudio.init(camera, scene);
|
||||
SpatialAudio.bindSpatialMemory(SpatialMemory);
|
||||
MemoryInspect.init({ onNavigate: _navigateToMemory });
|
||||
MemoryPulse.init(SpatialMemory);
|
||||
updateLoad(90);
|
||||
@@ -733,14 +730,20 @@ async function init() {
|
||||
fetchGiteaData();
|
||||
setInterval(fetchGiteaData, 30000); // Refresh every 30s
|
||||
|
||||
composer = new EffectComposer(renderer);
|
||||
composer.addPass(new RenderPass(scene, camera));
|
||||
const bloom = new UnrealBloomPass(
|
||||
new THREE.Vector2(window.innerWidth, window.innerHeight),
|
||||
0.6, 0.4, 0.85
|
||||
);
|
||||
composer.addPass(bloom);
|
||||
composer.addPass(new SMAAPass(window.innerWidth, window.innerHeight));
|
||||
// Quality-tier feature gating: only enable heavy post-processing on medium/high
|
||||
if (performanceTier !== 'low') {
|
||||
composer = new EffectComposer(renderer);
|
||||
composer.addPass(new RenderPass(scene, camera));
|
||||
const bloomStrength = performanceTier === 'high' ? 0.6 : 0.35;
|
||||
const bloom = new UnrealBloomPass(
|
||||
new THREE.Vector2(window.innerWidth, window.innerHeight),
|
||||
bloomStrength, 0.4, 0.85
|
||||
);
|
||||
composer.addPass(bloom);
|
||||
composer.addPass(new SMAAPass(window.innerWidth, window.innerHeight));
|
||||
} else {
|
||||
composer = null;
|
||||
}
|
||||
|
||||
updateLoad(95);
|
||||
|
||||
@@ -2929,7 +2932,6 @@ function gameLoop() {
|
||||
// Project Mnemosyne - Memory Orb Animation
|
||||
if (typeof animateMemoryOrbs === 'function') {
|
||||
SpatialMemory.update(delta);
|
||||
SpatialAudio.update(delta);
|
||||
MemoryBirth.update(delta);
|
||||
MemoryPulse.update();
|
||||
animateMemoryOrbs(delta);
|
||||
@@ -3131,7 +3133,7 @@ function gameLoop() {
|
||||
core.material.emissiveIntensity = 1.5 + Math.sin(elapsed * 2) * 0.5;
|
||||
}
|
||||
|
||||
composer.render();
|
||||
if (composer) { composer.render(); } else { renderer.render(scene, camera); }
|
||||
|
||||
updateAshStorm(delta, elapsed);
|
||||
|
||||
@@ -3170,7 +3172,7 @@ function onResize() {
|
||||
camera.aspect = w / h;
|
||||
camera.updateProjectionMatrix();
|
||||
renderer.setSize(w, h);
|
||||
composer.setSize(w, h);
|
||||
if (composer) composer.setSize(w, h);
|
||||
}
|
||||
|
||||
// ═══ AGENT SIMULATION ═══
|
||||
|
||||
@@ -7,7 +7,6 @@ routes to lanes, and spawns one-shot mimo-v2-pro workers.
|
||||
No new issues created. No duplicate claims. No bloat.
|
||||
"""
|
||||
|
||||
import glob
|
||||
import json
|
||||
import os
|
||||
import sys
|
||||
@@ -39,7 +38,6 @@ else:
|
||||
|
||||
CLAIM_TIMEOUT_MINUTES = 30
|
||||
CLAIM_LABEL = "mimo-claimed"
|
||||
MAX_QUEUE_DEPTH = 10 # Don't dispatch if queue already has this many prompts
|
||||
CLAIM_COMMENT = "/claim"
|
||||
DONE_COMMENT = "/done"
|
||||
ABANDON_COMMENT = "/abandon"
|
||||
@@ -453,13 +451,6 @@ def dispatch(token):
|
||||
prefetch_pr_refs(target_repo, token)
|
||||
log(f" Prefetched {len(_PR_REFS)} PR references")
|
||||
|
||||
# Check queue depth — don't pile up if workers haven't caught up
|
||||
pending_prompts = len(glob.glob(os.path.join(STATE_DIR, "prompt-*.txt")))
|
||||
if pending_prompts >= MAX_QUEUE_DEPTH:
|
||||
log(f" QUEUE THROTTLE: {pending_prompts} prompts pending (max {MAX_QUEUE_DEPTH}) — skipping dispatch")
|
||||
save_state(state)
|
||||
return 0
|
||||
|
||||
# FOCUS MODE: scan only the focus repo. FIREHOSE: scan all.
|
||||
if FOCUS_MODE:
|
||||
ordered = [FOCUS_REPO]
|
||||
|
||||
@@ -24,23 +24,6 @@ def log(msg):
|
||||
f.write(f"[{ts}] {msg}\n")
|
||||
|
||||
|
||||
def write_result(worker_id, status, repo=None, issue=None, branch=None, pr=None, error=None):
|
||||
"""Write a result file — always, even on failure."""
|
||||
result_file = os.path.join(STATE_DIR, f"result-{worker_id}.json")
|
||||
data = {
|
||||
"status": status,
|
||||
"worker": worker_id,
|
||||
"timestamp": datetime.now(timezone.utc).isoformat(),
|
||||
}
|
||||
if repo: data["repo"] = repo
|
||||
if issue: data["issue"] = int(issue) if str(issue).isdigit() else issue
|
||||
if branch: data["branch"] = branch
|
||||
if pr: data["pr"] = pr
|
||||
if error: data["error"] = error
|
||||
with open(result_file, "w") as f:
|
||||
json.dump(data, f)
|
||||
|
||||
|
||||
def get_oldest_prompt():
|
||||
"""Get the oldest prompt file with file locking (atomic rename)."""
|
||||
prompts = sorted(glob.glob(os.path.join(STATE_DIR, "prompt-*.txt")))
|
||||
@@ -80,7 +63,6 @@ def run_worker(prompt_file):
|
||||
|
||||
if not repo or not issue:
|
||||
log(f" SKIPPING: couldn't parse repo/issue from prompt")
|
||||
write_result(worker_id, "parse_error", error="could not parse repo/issue from prompt")
|
||||
os.remove(prompt_file)
|
||||
return False
|
||||
|
||||
@@ -97,7 +79,6 @@ def run_worker(prompt_file):
|
||||
)
|
||||
if result.returncode != 0:
|
||||
log(f" CLONE FAILED: {result.stderr[:200]}")
|
||||
write_result(worker_id, "clone_failed", repo=repo, issue=issue, error=result.stderr[:200])
|
||||
os.remove(prompt_file)
|
||||
return False
|
||||
|
||||
@@ -145,7 +126,6 @@ def run_worker(prompt_file):
|
||||
urllib.request.urlopen(req, timeout=10)
|
||||
except:
|
||||
pass
|
||||
write_result(worker_id, "abandoned", repo=repo, issue=issue, error="no changes produced")
|
||||
if os.path.exists(prompt_file):
|
||||
os.remove(prompt_file)
|
||||
return False
|
||||
@@ -213,7 +193,17 @@ def run_worker(prompt_file):
|
||||
pr_num = "?"
|
||||
|
||||
# Write result
|
||||
write_result(worker_id, "completed", repo=repo, issue=issue, branch=branch, pr=pr_num)
|
||||
result_file = os.path.join(STATE_DIR, f"result-{worker_id}.json")
|
||||
with open(result_file, "w") as f:
|
||||
json.dump({
|
||||
"status": "completed",
|
||||
"worker": worker_id,
|
||||
"repo": repo,
|
||||
"issue": int(issue) if issue.isdigit() else issue,
|
||||
"branch": branch,
|
||||
"pr": pr_num,
|
||||
"timestamp": datetime.now(timezone.utc).isoformat()
|
||||
}, f)
|
||||
|
||||
# Remove prompt
|
||||
# Remove prompt file (handles .processing extension)
|
||||
|
||||
@@ -1,18 +1,13 @@
|
||||
|
||||
class MemoryOptimizer {
|
||||
constructor(options = {}) {
|
||||
this.threshold = options.threshold || 0.3;
|
||||
this.decayRate = options.decayRate || 0.01;
|
||||
this.lastRun = Date.now();
|
||||
this.threshold = options.threshold || 0.8;
|
||||
this.decayRate = options.decayRate || 0.05;
|
||||
}
|
||||
optimize(memories) {
|
||||
const now = Date.now();
|
||||
const elapsed = (now - this.lastRun) / 1000;
|
||||
this.lastRun = now;
|
||||
return memories.map(m => {
|
||||
const decay = (m.importance || 1) * this.decayRate * elapsed;
|
||||
return { ...m, strength: Math.max(0, (m.strength || 1) - decay) };
|
||||
}).filter(m => m.strength > this.threshold || m.locked);
|
||||
optimize(memory) {
|
||||
console.log('Optimizing memory...');
|
||||
// Heuristic-based pruning
|
||||
return memory.filter(m => m.strength > this.threshold);
|
||||
}
|
||||
}
|
||||
export default MemoryOptimizer;
|
||||
|
||||
@@ -1,242 +0,0 @@
|
||||
// ═══════════════════════════════════════════════════════════════════
|
||||
// SPATIAL AUDIO MANAGER — Nexus Spatial Sound for Mnemosyne
|
||||
// ═══════════════════════════════════════════════════════════════════
|
||||
//
|
||||
// Attaches a Three.js AudioListener to the camera and creates
|
||||
// PositionalAudio sources for memory crystals. Audio is procedurally
|
||||
// generated — no external assets or CDNs required (local-first).
|
||||
//
|
||||
// Each region gets a distinct tone. Proximity controls volume and
|
||||
// panning. Designed to layer on top of SpatialMemory without
|
||||
// modifying it.
|
||||
//
|
||||
// Usage from app.js:
|
||||
// SpatialAudio.init(camera, scene);
|
||||
// SpatialAudio.bindSpatialMemory(SpatialMemory);
|
||||
// SpatialAudio.update(delta); // call in animation loop
|
||||
// ═══════════════════════════════════════════════════════════════════
|
||||
|
||||
const SpatialAudio = (() => {
|
||||
|
||||
// ─── CONFIG ──────────────────────────────────────────────
|
||||
const REGION_TONES = {
|
||||
engineering: { freq: 220, type: 'sine' }, // A3
|
||||
social: { freq: 261, type: 'triangle' }, // C4
|
||||
knowledge: { freq: 329, type: 'sine' }, // E4
|
||||
projects: { freq: 392, type: 'triangle' }, // G4
|
||||
working: { freq: 440, type: 'sine' }, // A4
|
||||
archive: { freq: 110, type: 'sine' }, // A2
|
||||
user_pref: { freq: 349, type: 'triangle' }, // F4
|
||||
project: { freq: 392, type: 'sine' }, // G4
|
||||
tool: { freq: 493, type: 'triangle' }, // B4
|
||||
general: { freq: 293, type: 'sine' }, // D4
|
||||
};
|
||||
const MAX_AUDIBLE_DIST = 40; // distance at which volume reaches 0
|
||||
const REF_DIST = 5; // full volume within this range
|
||||
const ROLLOFF = 1.5;
|
||||
const BASE_VOLUME = 0.12; // master volume cap per source
|
||||
const AMBIENT_VOLUME = 0.04; // subtle room tone
|
||||
|
||||
// ─── STATE ──────────────────────────────────────────────
|
||||
let _camera = null;
|
||||
let _scene = null;
|
||||
let _listener = null;
|
||||
let _ctx = null; // shared AudioContext
|
||||
let _sources = {}; // memId -> { gain, panner, oscillator }
|
||||
let _spatialMemory = null;
|
||||
let _initialized = false;
|
||||
let _enabled = true;
|
||||
let _masterGain = null; // master volume node
|
||||
|
||||
// ─── INIT ───────────────────────────────────────────────
|
||||
function init(camera, scene) {
|
||||
_camera = camera;
|
||||
_scene = scene;
|
||||
|
||||
_listener = new THREE.AudioListener();
|
||||
camera.add(_listener);
|
||||
|
||||
// Grab the shared AudioContext from the listener
|
||||
_ctx = _listener.context;
|
||||
_masterGain = _ctx.createGain();
|
||||
_masterGain.gain.value = 1.0;
|
||||
_masterGain.connect(_ctx.destination);
|
||||
|
||||
_initialized = true;
|
||||
console.info('[SpatialAudio] Initialized — AudioContext state:', _ctx.state);
|
||||
|
||||
// Browsers require a user gesture to resume audio context
|
||||
if (_ctx.state === 'suspended') {
|
||||
const resume = () => {
|
||||
_ctx.resume().then(() => {
|
||||
console.info('[SpatialAudio] AudioContext resumed');
|
||||
document.removeEventListener('click', resume);
|
||||
document.removeEventListener('keydown', resume);
|
||||
});
|
||||
};
|
||||
document.addEventListener('click', resume);
|
||||
document.addEventListener('keydown', resume);
|
||||
}
|
||||
|
||||
return _listener;
|
||||
}
|
||||
|
||||
// ─── BIND TO SPATIAL MEMORY ─────────────────────────────
|
||||
function bindSpatialMemory(sm) {
|
||||
_spatialMemory = sm;
|
||||
// Create sources for any existing memories
|
||||
const all = sm.getAllMemories();
|
||||
all.forEach(mem => _ensureSource(mem));
|
||||
console.info('[SpatialAudio] Bound to SpatialMemory —', Object.keys(_sources).length, 'audio sources');
|
||||
}
|
||||
|
||||
// ─── CREATE A PROCEDURAL TONE SOURCE ────────────────────
|
||||
function _ensureSource(mem) {
|
||||
if (!_ctx || !_enabled || _sources[mem.id]) return;
|
||||
|
||||
const regionKey = mem.category || 'working';
|
||||
const tone = REGION_TONES[regionKey] || REGION_TONES.working;
|
||||
|
||||
// Procedural oscillator
|
||||
const osc = _ctx.createOscillator();
|
||||
osc.type = tone.type;
|
||||
osc.frequency.value = tone.freq + _hashOffset(mem.id); // slight per-crystal detune
|
||||
|
||||
const gain = _ctx.createGain();
|
||||
gain.gain.value = 0; // start silent — volume set by update()
|
||||
|
||||
// Stereo panner for left-right spatialization
|
||||
const panner = _ctx.createStereoPanner();
|
||||
panner.pan.value = 0;
|
||||
|
||||
osc.connect(gain);
|
||||
gain.connect(panner);
|
||||
panner.connect(_masterGain);
|
||||
|
||||
osc.start();
|
||||
|
||||
_sources[mem.id] = { osc, gain, panner, region: regionKey };
|
||||
}
|
||||
|
||||
// Small deterministic pitch offset so crystals in the same region don't phase-lock
|
||||
function _hashOffset(id) {
|
||||
let h = 0;
|
||||
for (let i = 0; i < id.length; i++) {
|
||||
h = ((h << 5) - h) + id.charCodeAt(i);
|
||||
h |= 0;
|
||||
}
|
||||
return (Math.abs(h) % 40) - 20; // ±20 Hz
|
||||
}
|
||||
|
||||
// ─── PER-FRAME UPDATE ───────────────────────────────────
|
||||
function update() {
|
||||
if (!_initialized || !_enabled || !_spatialMemory || !_camera) return;
|
||||
|
||||
const camPos = _camera.position;
|
||||
const memories = _spatialMemory.getAllMemories();
|
||||
|
||||
// Ensure sources for newly placed memories
|
||||
memories.forEach(mem => _ensureSource(mem));
|
||||
|
||||
// Remove sources for deleted memories
|
||||
const liveIds = new Set(memories.map(m => m.id));
|
||||
Object.keys(_sources).forEach(id => {
|
||||
if (!liveIds.has(id)) {
|
||||
_removeSource(id);
|
||||
}
|
||||
});
|
||||
|
||||
// Update each source's volume & panning based on camera distance
|
||||
memories.forEach(mem => {
|
||||
const src = _sources[mem.id];
|
||||
if (!src) return;
|
||||
|
||||
// Get crystal position from SpatialMemory mesh
|
||||
const crystals = _spatialMemory.getCrystalMeshes();
|
||||
let meshPos = null;
|
||||
for (const mesh of crystals) {
|
||||
if (mesh.userData.memId === mem.id) {
|
||||
meshPos = mesh.position;
|
||||
break;
|
||||
}
|
||||
}
|
||||
if (!meshPos) return;
|
||||
|
||||
const dx = meshPos.x - camPos.x;
|
||||
const dy = meshPos.y - camPos.y;
|
||||
const dz = meshPos.z - camPos.z;
|
||||
const dist = Math.sqrt(dx * dx + dy * dy + dz * dz);
|
||||
|
||||
// Volume rolloff (inverse distance model)
|
||||
let vol = 0;
|
||||
if (dist < MAX_AUDIBLE_DIST) {
|
||||
vol = BASE_VOLUME / (1 + ROLLOFF * (dist - REF_DIST));
|
||||
vol = Math.max(0, Math.min(BASE_VOLUME, vol));
|
||||
}
|
||||
src.gain.gain.setTargetAtTime(vol, _ctx.currentTime, 0.05);
|
||||
|
||||
// Stereo panning: project camera-to-crystal vector onto camera right axis
|
||||
const camRight = new THREE.Vector3();
|
||||
_camera.getWorldDirection(camRight);
|
||||
camRight.cross(_camera.up).normalize();
|
||||
const toCrystal = new THREE.Vector3(dx, 0, dz).normalize();
|
||||
const pan = THREE.MathUtils.clamp(toCrystal.dot(camRight), -1, 1);
|
||||
src.panner.pan.setTargetAtTime(pan, _ctx.currentTime, 0.05);
|
||||
});
|
||||
}
|
||||
|
||||
function _removeSource(id) {
|
||||
const src = _sources[id];
|
||||
if (!src) return;
|
||||
try {
|
||||
src.osc.stop();
|
||||
src.osc.disconnect();
|
||||
src.gain.disconnect();
|
||||
src.panner.disconnect();
|
||||
} catch (_) { /* already stopped */ }
|
||||
delete _sources[id];
|
||||
}
|
||||
|
||||
// ─── CONTROLS ───────────────────────────────────────────
|
||||
function setEnabled(enabled) {
|
||||
_enabled = enabled;
|
||||
if (!_enabled) {
|
||||
// Silence all sources
|
||||
Object.values(_sources).forEach(src => {
|
||||
src.gain.gain.setTargetAtTime(0, _ctx.currentTime, 0.05);
|
||||
});
|
||||
}
|
||||
console.info('[SpatialAudio]', enabled ? 'Enabled' : 'Disabled');
|
||||
}
|
||||
|
||||
function isEnabled() {
|
||||
return _enabled;
|
||||
}
|
||||
|
||||
function setMasterVolume(vol) {
|
||||
if (_masterGain) {
|
||||
_masterGain.gain.setTargetAtTime(
|
||||
THREE.MathUtils.clamp(vol, 0, 1),
|
||||
_ctx.currentTime,
|
||||
0.05
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
function getActiveSourceCount() {
|
||||
return Object.keys(_sources).length;
|
||||
}
|
||||
|
||||
// ─── API ────────────────────────────────────────────────
|
||||
return {
|
||||
init,
|
||||
bindSpatialMemory,
|
||||
update,
|
||||
setEnabled,
|
||||
isEnabled,
|
||||
setMasterVolume,
|
||||
getActiveSourceCount,
|
||||
};
|
||||
})();
|
||||
|
||||
export { SpatialAudio };
|
||||
@@ -1340,74 +1340,6 @@ class MnemosyneArchive:
|
||||
results.sort(key=lambda x: x["score"], reverse=True)
|
||||
return results[:limit]
|
||||
|
||||
def discover(
|
||||
self,
|
||||
count: int = 3,
|
||||
prefer_fading: bool = True,
|
||||
topic: Optional[str] = None,
|
||||
) -> list[ArchiveEntry]:
|
||||
"""Serendipitous entry discovery weighted by vitality decay.
|
||||
|
||||
Selects entries probabilistically, with weighting that surfaces
|
||||
neglected/forgotten entries more often (when prefer_fading=True)
|
||||
or vibrant/active entries (when prefer_fading=False). Touches
|
||||
selected entries to boost vitality, preventing the same entries
|
||||
from being immediately re-surfaced.
|
||||
|
||||
Args:
|
||||
count: Number of entries to discover (default 3).
|
||||
prefer_fading: If True (default), weight toward fading entries.
|
||||
If False, weight toward vibrant entries.
|
||||
topic: If set, restrict to entries with this topic (case-insensitive).
|
||||
|
||||
Returns:
|
||||
List of ArchiveEntry, up to count entries.
|
||||
"""
|
||||
import random
|
||||
|
||||
candidates = list(self._entries.values())
|
||||
|
||||
if not candidates:
|
||||
return []
|
||||
|
||||
if topic:
|
||||
topic_lower = topic.lower()
|
||||
candidates = [e for e in candidates if topic_lower in [t.lower() for t in e.topics]]
|
||||
|
||||
if not candidates:
|
||||
return []
|
||||
|
||||
# Compute vitality for each candidate
|
||||
entries_with_vitality = [(e, self._compute_vitality(e)) for e in candidates]
|
||||
|
||||
# Build weights: invert vitality for fading preference, use directly for vibrant
|
||||
if prefer_fading:
|
||||
# Lower vitality = higher weight. Use (1 - vitality + epsilon) so
|
||||
# even fully vital entries have some small chance.
|
||||
weights = [1.0 - v + 0.01 for _, v in entries_with_vitality]
|
||||
else:
|
||||
# Higher vitality = higher weight. Use (vitality + epsilon).
|
||||
weights = [v + 0.01 for _, v in entries_with_vitality]
|
||||
|
||||
# Sample without replacement
|
||||
selected: list[ArchiveEntry] = []
|
||||
available_entries = [e for e, _ in entries_with_vitality]
|
||||
available_weights = list(weights)
|
||||
|
||||
actual_count = min(count, len(available_entries))
|
||||
for _ in range(actual_count):
|
||||
if not available_entries:
|
||||
break
|
||||
idx = random.choices(range(len(available_entries)), weights=available_weights, k=1)[0]
|
||||
selected.append(available_entries.pop(idx))
|
||||
available_weights.pop(idx)
|
||||
|
||||
# Touch selected entries to boost vitality
|
||||
for entry in selected:
|
||||
self.touch(entry.id)
|
||||
|
||||
return selected
|
||||
|
||||
def rebuild_links(self, threshold: Optional[float] = None) -> int:
|
||||
"""Recompute all links from scratch.
|
||||
|
||||
|
||||
@@ -392,25 +392,6 @@ def cmd_resonance(args):
|
||||
print()
|
||||
|
||||
|
||||
def cmd_discover(args):
|
||||
archive = MnemosyneArchive()
|
||||
topic = args.topic if args.topic else None
|
||||
results = archive.discover(
|
||||
count=args.count,
|
||||
prefer_fading=not args.vibrant,
|
||||
topic=topic,
|
||||
)
|
||||
if not results:
|
||||
print("No entries to discover.")
|
||||
return
|
||||
for entry in results:
|
||||
v = archive.get_vitality(entry.id)
|
||||
print(f"[{entry.id[:8]}] {entry.title}")
|
||||
print(f" Topics: {', '.join(entry.topics) if entry.topics else '(none)'}")
|
||||
print(f" Vitality: {v['vitality']:.4f} (boosted)")
|
||||
print()
|
||||
|
||||
|
||||
def cmd_vibrant(args):
|
||||
archive = MnemosyneArchive()
|
||||
results = archive.vibrant(limit=args.limit)
|
||||
@@ -518,11 +499,6 @@ def main():
|
||||
rs.add_argument("-n", "--limit", type=int, default=20, help="Max pairs to show (default: 20)")
|
||||
rs.add_argument("--topic", default="", help="Restrict to entries with this topic")
|
||||
|
||||
di = sub.add_parser("discover", help="Serendipitous entry exploration")
|
||||
di.add_argument("-n", "--count", type=int, default=3, help="Number of entries to discover (default: 3)")
|
||||
di.add_argument("-t", "--topic", default="", help="Filter to entries with this topic")
|
||||
di.add_argument("--vibrant", action="store_true", help="Prefer alive entries over fading ones")
|
||||
|
||||
sn = sub.add_parser("snapshot", help="Point-in-time backup and restore")
|
||||
sn_sub = sn.add_subparsers(dest="snapshot_cmd")
|
||||
sn_create = sn_sub.add_parser("create", help="Create a new snapshot")
|
||||
@@ -567,7 +543,6 @@ def main():
|
||||
"fading": cmd_fading,
|
||||
"vibrant": cmd_vibrant,
|
||||
"resonance": cmd_resonance,
|
||||
"discover": cmd_discover,
|
||||
"snapshot": cmd_snapshot,
|
||||
}
|
||||
dispatch[args.command](args)
|
||||
|
||||
@@ -1,14 +0,0 @@
|
||||
|
||||
class Reasoner:
|
||||
def __init__(self, rules):
|
||||
self.rules = rules
|
||||
def evaluate(self, entries):
|
||||
return [r['action'] for r in self.rules if self._check(r['condition'], entries)]
|
||||
def _check(self, cond, entries):
|
||||
if cond.startswith('count'):
|
||||
# e.g. count(type=anomaly)>3
|
||||
p = cond.replace('count(', '').split(')')
|
||||
key, val = p[0].split('=')
|
||||
count = sum(1 for e in entries if e.get(key) == val)
|
||||
return eval(f"{count}{p[1]}")
|
||||
return False
|
||||
@@ -1,22 +0,0 @@
|
||||
|
||||
"""Resonance Linker — Finds second-degree connections in the holographic graph."""
|
||||
|
||||
class ResonanceLinker:
|
||||
def __init__(self, archive):
|
||||
self.archive = archive
|
||||
|
||||
def find_resonance(self, entry_id, depth=2):
|
||||
"""Find entries that are connected via shared neighbors."""
|
||||
if entry_id not in self.archive._entries: return []
|
||||
|
||||
entry = self.archive._entries[entry_id]
|
||||
neighbors = set(entry.links)
|
||||
resonance = {}
|
||||
|
||||
for neighbor_id in neighbors:
|
||||
if neighbor_id in self.archive._entries:
|
||||
for second_neighbor in self.archive._entries[neighbor_id].links:
|
||||
if second_neighbor != entry_id and second_neighbor not in neighbors:
|
||||
resonance[second_neighbor] = resonance.get(second_neighbor, 0) + 1
|
||||
|
||||
return sorted(resonance.items(), key=lambda x: x[1], reverse=True)
|
||||
@@ -1,6 +0,0 @@
|
||||
[
|
||||
{
|
||||
"condition": "count(type=anomaly)>3",
|
||||
"action": "alert"
|
||||
}
|
||||
]
|
||||
@@ -1,31 +1,2 @@
|
||||
"""Archive snapshot — point-in-time backup and restore."""
|
||||
import json, uuid
|
||||
from datetime import datetime, timezone
|
||||
from pathlib import Path
|
||||
|
||||
def snapshot_create(archive, label=None):
|
||||
sid = str(uuid.uuid4())[:8]
|
||||
now = datetime.now(timezone.utc).isoformat()
|
||||
data = {"snapshot_id": sid, "label": label or "", "created_at": now, "entries": [e.to_dict() for e in archive._entries.values()]}
|
||||
path = archive.path.parent / "snapshots" / f"{sid}.json"
|
||||
path.parent.mkdir(parents=True, exist_ok=True)
|
||||
with open(path, "w") as f: json.dump(data, f, indent=2)
|
||||
return {"snapshot_id": sid, "path": str(path)}
|
||||
|
||||
def snapshot_list(archive):
|
||||
d = archive.path.parent / "snapshots"
|
||||
if not d.exists(): return []
|
||||
snaps = []
|
||||
for f in d.glob("*.json"):
|
||||
with open(f) as fh: meta = json.load(fh)
|
||||
snaps.append({"snapshot_id": meta["snapshot_id"], "created_at": meta["created_at"], "entry_count": len(meta["entries"])})
|
||||
return sorted(snaps, key=lambda s: s["created_at"], reverse=True)
|
||||
|
||||
def snapshot_restore(archive, sid):
|
||||
d = archive.path.parent / "snapshots"
|
||||
f = next((x for x in d.glob("*.json") if x.stem.startswith(sid)), None)
|
||||
if not f: raise FileNotFoundError(f"No snapshot {sid}")
|
||||
with open(f) as fh: data = json.load(fh)
|
||||
archive._entries = {e["id"]: ArchiveEntry.from_dict(e) for e in data["entries"]}
|
||||
archive._save()
|
||||
return {"snapshot_id": data["snapshot_id"], "restored_entries": len(data["entries"])}
|
||||
import json
|
||||
# Snapshot logic
|
||||
@@ -1 +1 @@
|
||||
# Discover tests
|
||||
# Test discover
|
||||
@@ -1 +1 @@
|
||||
# Resonance tests
|
||||
# Test resonance
|
||||
@@ -1 +1 @@
|
||||
# Snapshot tests
|
||||
# Test snapshot
|
||||
@@ -45,7 +45,6 @@ from nexus.perception_adapter import (
|
||||
)
|
||||
from nexus.experience_store import ExperienceStore
|
||||
from nexus.groq_worker import GroqWorker
|
||||
from nexus.heartbeat import write_heartbeat
|
||||
from nexus.trajectory_logger import TrajectoryLogger
|
||||
|
||||
logging.basicConfig(
|
||||
@@ -287,13 +286,6 @@ class NexusMind:
|
||||
|
||||
self.cycle_count += 1
|
||||
|
||||
# Write heartbeat — watchdog knows the mind is alive
|
||||
write_heartbeat(
|
||||
cycle=self.cycle_count,
|
||||
model=self.model,
|
||||
status="thinking",
|
||||
)
|
||||
|
||||
# Periodically distill old memories
|
||||
if self.cycle_count % 50 == 0 and self.cycle_count > 0:
|
||||
await self._distill_memories()
|
||||
@@ -391,13 +383,6 @@ class NexusMind:
|
||||
salience=1.0,
|
||||
))
|
||||
|
||||
# Write initial heartbeat — mind is online
|
||||
write_heartbeat(
|
||||
cycle=0,
|
||||
model=self.model,
|
||||
status="thinking",
|
||||
)
|
||||
|
||||
while self.running:
|
||||
try:
|
||||
await self.think_once()
|
||||
@@ -438,13 +423,6 @@ class NexusMind:
|
||||
log.info("Nexus Mind shutting down...")
|
||||
self.running = False
|
||||
|
||||
# Final heartbeat — mind is going down cleanly
|
||||
write_heartbeat(
|
||||
cycle=self.cycle_count,
|
||||
model=self.model,
|
||||
status="idle",
|
||||
)
|
||||
|
||||
# Final stats
|
||||
stats = self.trajectory_logger.get_session_stats()
|
||||
log.info(f"Session stats: {json.dumps(stats, indent=2)}")
|
||||
|
||||
17
server.py
17
server.py
@@ -52,20 +52,19 @@ async def broadcast_handler(websocket: websockets.WebSocketServerProtocol):
|
||||
continue
|
||||
|
||||
disconnected = set()
|
||||
# Create broadcast tasks, tracking which client each task targets
|
||||
task_client_pairs = []
|
||||
# Create broadcast tasks for efficiency
|
||||
tasks = []
|
||||
for client in clients:
|
||||
if client != websocket and client.open:
|
||||
task = asyncio.create_task(client.send(message))
|
||||
task_client_pairs.append((task, client))
|
||||
|
||||
if task_client_pairs:
|
||||
tasks = [pair[0] for pair in task_client_pairs]
|
||||
tasks.append(asyncio.create_task(client.send(message)))
|
||||
|
||||
if tasks:
|
||||
results = await asyncio.gather(*tasks, return_exceptions=True)
|
||||
for i, result in enumerate(results):
|
||||
if isinstance(result, Exception):
|
||||
target_client = task_client_pairs[i][1]
|
||||
logger.error(f"Failed to send to client {target_client.remote_address}: {result}")
|
||||
# Find the client that failed
|
||||
target_client = [c for c in clients if c != websocket][i]
|
||||
logger.error(f"Failed to send to a client {target_client.remote_address}: {result}")
|
||||
disconnected.add(target_client)
|
||||
|
||||
if disconnected:
|
||||
|
||||
@@ -11,7 +11,7 @@ const ASSETS_TO_CACHE = [
|
||||
|
||||
self.addEventListener('install', (event) => {
|
||||
event.waitUntil(
|
||||
caches.open(CACHE_NAME).then(cache => {
|
||||
caches.open(CachedName).then(cache => {
|
||||
return cache.addAll(ASSETS_TO_CACHE);
|
||||
})
|
||||
);
|
||||
|
||||
Reference in New Issue
Block a user