Compare commits
58 Commits
feat/mnemo
...
mimo/code/
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
0a0a2eb802 | ||
| 6786e65f3d | |||
| 62a6581827 | |||
| 797f32a7fe | |||
| 80eb4ff7ea | |||
| b205f002ef | |||
| 2230c1c9fc | |||
| d7bcadb8c1 | |||
| e939958f38 | |||
| 387084e27f | |||
| 2661a9991f | |||
| a9604cbd7b | |||
| a16c2445ab | |||
| 36db3aff6b | |||
| 43f3da8e7d | |||
| 6e97542ebc | |||
| 6aafc7cbb8 | |||
| 84121936f0 | |||
| ba18e5ed5f | |||
| c3ae479661 | |||
| 9e04030541 | |||
| 75f11b4f48 | |||
| 72d9c1a303 | |||
| fd8f82315c | |||
| bb21beccdd | |||
| 3361a0e259 | |||
| 8fb0a50b91 | |||
| 99e4baf54b | |||
| b0e24af7fe | |||
| 65cef9d9c0 | |||
| 267505a68f | |||
| e8312d91f7 | |||
| 446ec370c8 | |||
| 76e62fe43f | |||
| b52c7281f0 | |||
| af1221fb80 | |||
| 42a4169940 | |||
| 3f7c037562 | |||
| 17e714c9d2 | |||
| 653c20862c | |||
| 89e19dbaa2 | |||
| 3fca28b1c8 | |||
| 1f8994abc9 | |||
| fcdb049117 | |||
| 85dda06ff0 | |||
| bd27cd4bf5 | |||
| fd7c66bd54 | |||
| 3bf8d6e0a6 | |||
| eeba35b3a9 | |||
|
|
55f0bbe97e | ||
|
|
410cd12172 | ||
|
|
abe8c9f790 | ||
|
|
67adf79757 | ||
| a378aa576e | |||
|
|
5446d3dc59 | ||
|
|
58c75a29bd | ||
| b3939179b9 | |||
| a14bf80631 |
201
.githooks/stale-pr-closer.sh
Executable file
201
.githooks/stale-pr-closer.sh
Executable file
@@ -0,0 +1,201 @@
|
||||
#!/usr/bin/env bash
|
||||
# ═══════════════════════════════════════════════════════════════
|
||||
# stale-pr-closer.sh — Auto-close conflicted PRs superseded by
|
||||
# already-merged work.
|
||||
#
|
||||
# Designed for cron on Hermes:
|
||||
# 0 */6 * * * /path/to/the-nexus/.githooks/stale-pr-closer.sh
|
||||
#
|
||||
# Closes #1250 (parent epic #1248)
|
||||
# ═══════════════════════════════════════════════════════════════
|
||||
set -euo pipefail
|
||||
|
||||
# ─── Configuration ──────────────────────────────────────────
|
||||
GITEA_URL="${GITEA_URL:-https://forge.alexanderwhitestone.com}"
|
||||
GITEA_TOKEN="${GITEA_TOKEN:?Set GITEA_TOKEN env var}"
|
||||
REPO="${REPO:-Timmy_Foundation/the-nexus}"
|
||||
GRACE_HOURS="${GRACE_HOURS:-24}"
|
||||
DRY_RUN="${DRY_RUN:-false}"
|
||||
|
||||
API="$GITEA_URL/api/v1"
|
||||
AUTH="Authorization: token $GITEA_TOKEN"
|
||||
|
||||
log() { echo "[$(date -u +%Y-%m-%dT%H:%M:%SZ)] $*"; }
|
||||
|
||||
# ─── Fetch open PRs ────────────────────────────────────────
|
||||
log "Checking open PRs for $REPO (grace period: ${GRACE_HOURS}h, dry_run: $DRY_RUN)"
|
||||
|
||||
OPEN_PRS=$(curl -s -H "$AUTH" "$API/repos/$REPO/pulls?state=open&limit=50")
|
||||
PR_COUNT=$(echo "$OPEN_PRS" | python3 -c "import json,sys; print(len(json.loads(sys.stdin.read())))")
|
||||
|
||||
if [ "$PR_COUNT" = "0" ]; then
|
||||
log "No open PRs. Done."
|
||||
exit 0
|
||||
fi
|
||||
|
||||
log "Found $PR_COUNT open PR(s)"
|
||||
|
||||
# ─── Fetch recently merged PRs (for supersession check) ────
|
||||
MERGED_PRS=$(curl -s -H "$AUTH" "$API/repos/$REPO/pulls?state=closed&limit=100&sort=updated&direction=desc")
|
||||
|
||||
# ─── Process each open PR ──────────────────────────────────
|
||||
echo "$OPEN_PRS" | python3 -c "
|
||||
import json, sys, re
|
||||
from datetime import datetime, timedelta, timezone
|
||||
|
||||
grace_hours = int('$GRACE_HOURS')
|
||||
dry_run = '$DRY_RUN' == 'true'
|
||||
api = '$API'
|
||||
repo = '$REPO'
|
||||
|
||||
open_prs = json.loads(sys.stdin.read())
|
||||
|
||||
# Read merged PRs from file we'll pipe separately
|
||||
# (We handle this in the shell wrapper below)
|
||||
" 2>/dev/null || true
|
||||
|
||||
# Use Python for the complex logic
|
||||
python3 << 'PYEOF'
|
||||
import json, sys, os, re, subprocess
|
||||
from datetime import datetime, timedelta, timezone
|
||||
|
||||
GITEA_URL = os.environ.get("GITEA_URL", "https://forge.alexanderwhitestone.com")
|
||||
GITEA_TOKEN = os.environ["GITEA_TOKEN"]
|
||||
REPO = os.environ.get("REPO", "Timmy_Foundation/the-nexus")
|
||||
GRACE_HOURS = int(os.environ.get("GRACE_HOURS", "24"))
|
||||
DRY_RUN = os.environ.get("DRY_RUN", "false") == "true"
|
||||
|
||||
API = f"{GITEA_URL}/api/v1"
|
||||
HEADERS = {"Authorization": f"token {GITEA_TOKEN}", "Content-Type": "application/json"}
|
||||
|
||||
import urllib.request, urllib.error
|
||||
|
||||
def api_get(path):
|
||||
req = urllib.request.Request(f"{API}{path}", headers=HEADERS)
|
||||
with urllib.request.urlopen(req) as resp:
|
||||
return json.loads(resp.read())
|
||||
|
||||
def api_post(path, data):
|
||||
body = json.dumps(data).encode()
|
||||
req = urllib.request.Request(f"{API}{path}", data=body, headers=HEADERS, method="POST")
|
||||
with urllib.request.urlopen(req) as resp:
|
||||
return json.loads(resp.read())
|
||||
|
||||
def api_patch(path, data):
|
||||
body = json.dumps(data).encode()
|
||||
req = urllib.request.Request(f"{API}{path}", data=body, headers=HEADERS, method="PATCH")
|
||||
with urllib.request.urlopen(req) as resp:
|
||||
return json.loads(resp.read())
|
||||
|
||||
def log(msg):
|
||||
from datetime import datetime, timezone
|
||||
ts = datetime.now(timezone.utc).strftime("%Y-%m-%dT%H:%M:%SZ")
|
||||
print(f"[{ts}] {msg}")
|
||||
|
||||
now = datetime.now(timezone.utc)
|
||||
cutoff = now - timedelta(hours=GRACE_HOURS)
|
||||
|
||||
# Fetch open PRs
|
||||
open_prs = api_get(f"/repos/{REPO}/pulls?state=open&limit=50")
|
||||
if not open_prs:
|
||||
log("No open PRs. Done.")
|
||||
sys.exit(0)
|
||||
|
||||
log(f"Found {len(open_prs)} open PR(s)")
|
||||
|
||||
# Fetch recently merged PRs
|
||||
merged_prs = api_get(f"/repos/{REPO}/pulls?state=closed&limit=100&sort=updated&direction=desc")
|
||||
merged_prs = [p for p in merged_prs if p.get("merged")]
|
||||
|
||||
# Build lookup: issue_number -> merged PR that closes it
|
||||
# Parse "Closes #NNN" from merged PR bodies
|
||||
def extract_closes(body):
|
||||
if not body:
|
||||
return set()
|
||||
return set(int(m) for m in re.findall(r'(?:closes?|fixes?|resolves?)\s+#(\d+)', body, re.IGNORECASE))
|
||||
|
||||
merged_by_issue = {}
|
||||
for mp in merged_prs:
|
||||
for issue_num in extract_closes(mp.get("body", "")):
|
||||
merged_by_issue[issue_num] = mp
|
||||
|
||||
# Also build a lookup by title similarity (for PRs that implement same feature without referencing same issue)
|
||||
merged_by_title_words = {}
|
||||
for mp in merged_prs:
|
||||
# Extract meaningful words from title
|
||||
title = re.sub(r'\[claude\]|\[.*?\]|feat\(.*?\):', '', mp.get("title", "")).strip().lower()
|
||||
words = set(w for w in re.findall(r'\w+', title) if len(w) > 3)
|
||||
if words:
|
||||
merged_by_title_words[mp["number"]] = (words, mp)
|
||||
|
||||
closed_count = 0
|
||||
|
||||
for pr in open_prs:
|
||||
pr_num = pr["number"]
|
||||
pr_title = pr["title"]
|
||||
mergeable = pr.get("mergeable", True)
|
||||
updated_at = datetime.fromisoformat(pr["updated_at"].replace("Z", "+00:00"))
|
||||
|
||||
# Skip if within grace period
|
||||
if updated_at > cutoff:
|
||||
log(f" PR #{pr_num}: within grace period, skipping")
|
||||
continue
|
||||
|
||||
# Check 1: Is it conflicted?
|
||||
if mergeable:
|
||||
log(f" PR #{pr_num}: mergeable, skipping")
|
||||
continue
|
||||
|
||||
# Check 2: Does a merged PR close the same issue?
|
||||
pr_closes = extract_closes(pr.get("body", ""))
|
||||
superseded_by = None
|
||||
|
||||
for issue_num in pr_closes:
|
||||
if issue_num in merged_by_issue:
|
||||
superseded_by = merged_by_issue[issue_num]
|
||||
break
|
||||
|
||||
# Check 3: Title similarity match (if no issue match)
|
||||
if not superseded_by:
|
||||
pr_title_clean = re.sub(r'\[.*?\]|feat\(.*?\):', '', pr_title).strip().lower()
|
||||
pr_words = set(w for w in re.findall(r'\w+', pr_title_clean) if len(w) > 3)
|
||||
|
||||
best_overlap = 0
|
||||
for mp_num, (mp_words, mp) in merged_by_title_words.items():
|
||||
if mp_num == pr_num:
|
||||
continue
|
||||
overlap = len(pr_words & mp_words)
|
||||
# Require at least 60% word overlap
|
||||
if pr_words and overlap / len(pr_words) >= 0.6 and overlap > best_overlap:
|
||||
best_overlap = overlap
|
||||
superseded_by = mp
|
||||
|
||||
if not superseded_by:
|
||||
log(f" PR #{pr_num}: conflicted but no superseding PR found, skipping")
|
||||
continue
|
||||
|
||||
sup_num = superseded_by["number"]
|
||||
sup_title = superseded_by["title"]
|
||||
merged_at = superseded_by.get("merged_at", "unknown")[:10]
|
||||
|
||||
comment = (
|
||||
f"**Auto-closed by stale-pr-closer**\n\n"
|
||||
f"This PR has merge conflicts and has been superseded by #{sup_num} "
|
||||
f"(\"{sup_title}\"), merged {merged_at}.\n\n"
|
||||
f"If this PR contains unique work not covered by #{sup_num}, "
|
||||
f"please reopen and rebase against `main`."
|
||||
)
|
||||
|
||||
if DRY_RUN:
|
||||
log(f" [DRY RUN] Would close PR #{pr_num} — superseded by #{sup_num}")
|
||||
else:
|
||||
# Post comment
|
||||
api_post(f"/repos/{REPO}/issues/{pr_num}/comments", {"body": comment})
|
||||
# Close PR
|
||||
api_patch(f"/repos/{REPO}/pulls/{pr_num}", {"state": "closed"})
|
||||
log(f" Closed PR #{pr_num} — superseded by #{sup_num} ({sup_title})")
|
||||
|
||||
closed_count += 1
|
||||
|
||||
log(f"Done. {'Would close' if DRY_RUN else 'Closed'} {closed_count} stale PR(s).")
|
||||
PYEOF
|
||||
19
.gitignore
vendored
19
.gitignore
vendored
@@ -1,11 +1,18 @@
|
||||
# === Python bytecode (recursive — covers all subdirectories) ===
|
||||
**/__pycache__/
|
||||
*.pyc
|
||||
*.pyo
|
||||
|
||||
# === Node ===
|
||||
node_modules/
|
||||
|
||||
# === Test artifacts ===
|
||||
test-results/
|
||||
nexus/__pycache__/
|
||||
tests/__pycache__/
|
||||
mempalace/__pycache__/
|
||||
test-screenshots/
|
||||
|
||||
# === Tool configs ===
|
||||
.aider*
|
||||
|
||||
# Prevent agents from writing to wrong path (see issue #1145)
|
||||
# === Path guardrails (see issue #1145) ===
|
||||
# Prevent agents from writing to wrong path
|
||||
public/nexus/
|
||||
test-screenshots/
|
||||
__pycache__/
|
||||
|
||||
480
CONTRIBUTING.md
480
CONTRIBUTING.md
@@ -1,206 +1,54 @@
|
||||
# Contribution & Code Review Policy
|
||||
# Contributing to The Nexus
|
||||
|
||||
## Issue Assignment — The Lock Protocol
|
||||
|
||||
**Rule: Assign before you code.**
|
||||
|
||||
Before starting work on any issue, you **must** assign it to yourself. If an issue is already assigned to someone else, **do not submit a competing PR**.
|
||||
|
||||
### For Humans
|
||||
|
||||
1. Check the issue is unassigned
|
||||
2. Assign yourself via the Gitea UI (right sidebar → Assignees)
|
||||
3. Start coding
|
||||
|
||||
### For Agents (Claude, Perplexity, Mimo, etc.)
|
||||
|
||||
1. Before generating code, call the Gitea API to check assignment:
|
||||
```
|
||||
GET /api/v1/repos/{owner}/{repo}/issues/{number}
|
||||
→ Check assignees array
|
||||
```
|
||||
2. If unassigned, self-assign:
|
||||
```
|
||||
POST /api/v1/repos/{owner}/{repo}/issues/{number}/assignees
|
||||
{"assignees": ["your-username"]}
|
||||
```
|
||||
3. If already assigned, **stop**. Post a comment offering to help instead.
|
||||
|
||||
### Why This Matters
|
||||
|
||||
On April 11, 2026, we found 12 stale PRs caused by Rockachopa and the `[claude]` auto-bot racing on the same issues. The auto-bot merged first, orphaning the manual PRs. Assignment-as-lock prevents this race condition.
|
||||
|
||||
---
|
||||
|
||||
## Branch Protection & Review Policy
|
||||
|
||||
All repositories enforce these rules on the `main` branch:
|
||||
- ✅ Require Pull Request for merge
|
||||
- ✅ Require 1 approval before merge
|
||||
- ✅ Dismiss stale approvals on new commits
|
||||
- <20>️ Require CI to pass (where CI exists)
|
||||
- ✅ Block force pushes to `main`
|
||||
- ✅ Block deletion of `main` branch
|
||||
All repositories enforce these rules on `main`:
|
||||
|
||||
### Default Reviewer Assignments
|
||||
|
||||
| Repository | Required Reviewers |
|
||||
|------------------|---------------------------------|
|
||||
| `hermes-agent` | `@perplexity`, `@Timmy` |
|
||||
| `the-nexus` | `@perplexity` |
|
||||
| `timmy-home` | `@perplexity` |
|
||||
| `timmy-config` | `@perplexity` |
|
||||
|
||||
### CI Enforcement Status
|
||||
|
||||
| Repository | CI Status |
|
||||
|------------------|---------------------------------|
|
||||
| `hermes-agent` | ✅ Active |
|
||||
| `the-nexus` | <20>️ CI runner pending (#915) |
|
||||
| `timmy-home` | ❌ No CI |
|
||||
| `timmy-config` | ❌ Limited CI |
|
||||
|
||||
### Workflow Requirements
|
||||
|
||||
1. Create feature branch from `main`
|
||||
2. Submit PR with clear description
|
||||
3. Wait for @perplexity review
|
||||
4. Address feedback if any
|
||||
5. Merge after approval and passing CI
|
||||
|
||||
### Emergency Exceptions
|
||||
Hotfixes require:
|
||||
- ✅ @Timmy approval
|
||||
- ✅ Post-merge documentation
|
||||
- ✅ Follow-up PR for full review
|
||||
|
||||
### Abandoned PR Policy
|
||||
- PRs inactive >7 day: 🧹 archived
|
||||
- Unreviewed PRs >14 days: ❌ closed
|
||||
|
||||
### Policy Enforcement
|
||||
These rules are enforced by Gitea branch protection settings. Direct pushes to main will be blocked.
|
||||
- Require rebase to re-enable
|
||||
|
||||
## Enforcement
|
||||
|
||||
These rules are enforced by Gitea's branch protection settings. Violations will be blocked at the platform level.
|
||||
# Contribution and Code Review Policy
|
||||
|
||||
## Branch Protection Rules
|
||||
|
||||
All repositories must enforce the following rules on the `main` branch:
|
||||
- ✅ Require Pull Request for merge
|
||||
- ✅ Require 1 approval before merge
|
||||
- ✅ Dismiss stale approvals when new commits are pushed
|
||||
- ✅ Require status checks to pass (where CI is configured)
|
||||
- ✅ Block force-pushing to `main`
|
||||
- ✅ Block deleting the `main` branch
|
||||
|
||||
## Default Reviewer Assignment
|
||||
|
||||
All repositories must configure the following default reviewers:
|
||||
- `@perplexity` as default reviewer for all repositories
|
||||
- `@Timmy` as required reviewer for `hermes-agent`
|
||||
- Repo-specific owners for specialized areas
|
||||
|
||||
## Implementation Status
|
||||
|
||||
| Repository | Branch Protection | CI Enforcement | Default Reviewers |
|
||||
|------------------|------------------|----------------|-------------------|
|
||||
| hermes-agent | ✅ Enabled | ✅ Active | @perplexity, @Timmy |
|
||||
| the-nexus | ✅ Enabled | ⚠️ CI pending | @perplexity |
|
||||
| timmy-home | ✅ Enabled | ❌ No CI | @perplexity |
|
||||
| timmy-config | ✅ Enabled | ❌ No CI | @perplexity |
|
||||
|
||||
## Compliance Requirements
|
||||
|
||||
All contributors must:
|
||||
1. Never push directly to `main`
|
||||
2. Create a pull request for all changes
|
||||
3. Get at least one approval before merging
|
||||
4. Ensure CI passes before merging (where applicable)
|
||||
|
||||
## Policy Enforcement
|
||||
|
||||
This policy is enforced via Gitea branch protection rules. Violations will be blocked at the platform level.
|
||||
|
||||
For questions about this policy, contact @perplexity or @Timmy.
|
||||
|
||||
### Required for All Merges
|
||||
- [x] Pull Request must exist for all changes
|
||||
- [x] At least 1 approval from reviewer
|
||||
- [x] CI checks must pass (where applicable)
|
||||
- [x] No force pushes allowed
|
||||
- [x] No direct pushes to main
|
||||
- [x] No branch deletion
|
||||
|
||||
### Review Requirements
|
||||
- [x] @perplexity must be assigned as reviewer
|
||||
- [x] @Timmy must review all changes to `hermes-agent/`
|
||||
- [x] No self-approvals allowed
|
||||
|
||||
### CI/CD Enforcement
|
||||
- [x] CI must be configured for all new features
|
||||
- [x] Failing CI blocks merge
|
||||
- [x] CI status displayed in PR header
|
||||
|
||||
### Abandoned PR Policy
|
||||
- PRs inactive >7 days get "needs attention" label
|
||||
- PRs inactive >21 days are archived
|
||||
- PRs inactive >90 days are closed
|
||||
- [ ] At least 1 approval from reviewer
|
||||
- [ ] CI checks must pass (where available)
|
||||
- [ ] No force pushes allowed
|
||||
- [ ] No direct pushes to main
|
||||
- [ ] No branch deletion
|
||||
|
||||
### Review Requirements by Repository
|
||||
```yaml
|
||||
hermes-agent:
|
||||
required_owners:
|
||||
- perplexity
|
||||
- Timmy
|
||||
|
||||
the-nexus:
|
||||
required_owners:
|
||||
- perplexity
|
||||
|
||||
timmy-home:
|
||||
required_owners:
|
||||
- perplexity
|
||||
|
||||
timmy-config:
|
||||
required_owners:
|
||||
- perplexity
|
||||
```
|
||||
|
||||
### CI Status
|
||||
|
||||
```text
|
||||
- hermes-agent: ✅ Active
|
||||
- the-nexus: ⚠️ CI runner disabled (see #915)
|
||||
- timmy-home: - (No CI)
|
||||
- timmy-config: - (Limited CI)
|
||||
```
|
||||
|
||||
### Branch Protection Status
|
||||
|
||||
All repositories now enforce:
|
||||
- Require PR for merge
|
||||
- 1+ approvals required
|
||||
- CI/CD must pass (where applicable)
|
||||
- Force push and branch deletion blocked
|
||||
- hermes-agent: ✅ Active
|
||||
- the-nexus: ⚠️ CI runner disabled (see #915)
|
||||
- timmy-home: - (No CI)
|
||||
- timmy-config: - (Limited CI)
|
||||
```
|
||||
|
||||
## Workflow
|
||||
1. Create feature branch
|
||||
2. Open PR against main
|
||||
3. Get 1+ approvals
|
||||
4. Ensure CI passes
|
||||
5. Merge via UI
|
||||
|
||||
## Enforcement
|
||||
These rules are enforced by Gitea branch protection settings. Direct pushes to main will be blocked.
|
||||
|
||||
## Abandoned PRs
|
||||
PRs not updated in >7 days will be labeled "stale" and may be closed after 30 days of inactivity.
|
||||
# Contributing to the Nexus
|
||||
|
||||
**Every PR: net ≤ 10 added lines.** Not a guideline — a hard limit.
|
||||
Add 40, remove 30. Can't remove? You're homebrewing. Import instead.
|
||||
|
||||
## Branch Protection & Review Policy
|
||||
|
||||
### Branch Protection Rules
|
||||
|
||||
All repositories enforce the following rules on the `main` branch:
|
||||
|
||||
| Rule | Status | Applies To |
|
||||
|------|--------|------------|
|
||||
| Require Pull Request for merge | ✅ Enabled | All |
|
||||
| Require 1 approval before merge | ✅ Enabled | All |
|
||||
| Dismiss stale approvals on new commits | ✅ Enabled | All |
|
||||
| Require CI to pass (where CI exists) | ⚠️ Conditional | All |
|
||||
| Block force pushes to `main` | ✅ Enabled | All |
|
||||
| Block deletion of `main` branch | ✅ Enabled | All |
|
||||
| Rule | Status |
|
||||
|------|--------|
|
||||
| Require Pull Request for merge | ✅ Enabled |
|
||||
| Require 1 approval before merge | ✅ Enabled |
|
||||
| Dismiss stale approvals on new commits | ✅ Enabled |
|
||||
| Require CI to pass (where CI exists) | ⚠️ Conditional |
|
||||
| Block force pushes to `main` | ✅ Enabled |
|
||||
| Block deletion of `main` branch | ✅ Enabled |
|
||||
|
||||
### Default Reviewer Assignments
|
||||
|
||||
| Repository | Required Reviewers |
|
||||
|------------|------------------|
|
||||
|------------|-------------------|
|
||||
| `hermes-agent` | `@perplexity`, `@Timmy` |
|
||||
| `the-nexus` | `@perplexity` |
|
||||
| `timmy-home` | `@perplexity` |
|
||||
@@ -215,199 +63,93 @@ All repositories enforce the following rules on the `main` branch:
|
||||
| `timmy-home` | ❌ No CI |
|
||||
| `timmy-config` | ❌ Limited CI |
|
||||
|
||||
### Review Requirements
|
||||
---
|
||||
|
||||
- All PRs must be reviewed by at least one reviewer
|
||||
- `@perplexity` is the default reviewer for all repositories
|
||||
- `@Timmy` is a required reviewer for `hermes-agent`
|
||||
## Branch Naming
|
||||
|
||||
All repositories enforce:
|
||||
- ✅ Require Pull Request for merge
|
||||
- ✅ Require 1 approval
|
||||
- ⚠<> Require CI to pass (CI runner pending)
|
||||
- ✅ Dismiss stale approvals on new commits
|
||||
- ✅ Block force pushes
|
||||
- ✅ Block branch deletion
|
||||
Use descriptive prefixes:
|
||||
|
||||
## Review Requirements
|
||||
| Prefix | Use |
|
||||
|--------|-----|
|
||||
| `feat/` | New features |
|
||||
| `fix/` | Bug fixes |
|
||||
| `epic/` | Multi-issue epic branches |
|
||||
| `docs/` | Documentation only |
|
||||
|
||||
- Mandatory reviewer: `@perplexity` for all repos
|
||||
- Mandatory reviewer: `@Timmy` for `hermes-agent/`
|
||||
- Optional: Add repo-specific owners for specialized areas
|
||||
Example: `feat/mnemosyne-memory-decay`
|
||||
|
||||
## Implementation Status
|
||||
---
|
||||
|
||||
- ✅ hermes-agent: All protections enabled
|
||||
- ✅ the-nexus: PR + 1 approval enforced
|
||||
- ✅ timmy-home: PR + 1 approval enforced
|
||||
- ✅ timmy-config: PR + 1 approval enforced
|
||||
## PR Requirements
|
||||
|
||||
> CI enforcement pending runner restoration (#915)
|
||||
1. **Rebase before merge** — PRs must be up-to-date with `main`. If you have merge conflicts, rebase locally and force-push.
|
||||
2. **Reference the issue** — Use `Closes #NNN` in the PR body so Gitea auto-closes the issue on merge.
|
||||
3. **No bytecode** — Never commit `__pycache__/` or `.pyc` files. The `.gitignore` handles this, but double-check.
|
||||
4. **One feature per PR** — Avoid omnibus PRs that bundle multiple unrelated features. They're harder to review and more likely to conflict.
|
||||
|
||||
## What gets preserved from legacy Matrix
|
||||
---
|
||||
|
||||
High-value candidates include:
|
||||
- visitor movement / embodiment
|
||||
- chat, bark, and presence systems
|
||||
- transcript logging
|
||||
- ambient / visual atmosphere systems
|
||||
- economy / satflow visualizations
|
||||
- smoke and browser validation discipline
|
||||
## Path Conventions
|
||||
|
||||
Those
|
||||
```
|
||||
| Module | Canon Path |
|
||||
|--------|-----------|
|
||||
| Mnemosyne (backend) | `nexus/mnemosyne/` |
|
||||
| Mnemosyne (frontend) | `nexus/components/` |
|
||||
| MemPalace | `nexus/mempalace/` |
|
||||
| Scripts/tools | `bin/` |
|
||||
| Git hooks/automation | `.githooks/` |
|
||||
| Tests | `nexus/mnemosyne/tests/` |
|
||||
|
||||
README.md
|
||||
````
|
||||
<<<<<<< SEARCH
|
||||
# Contribution & Code Review Policy
|
||||
**Never** create a duplicate module at the repo root (e.g., `mnemosyne/` when `nexus/mnemosyne/` already exists). Check `FEATURES.yaml` manifests for the canonical path.
|
||||
|
||||
## Branch Protection Rules (Enforced via Gitea)
|
||||
All repositories must have the following branch protection rules enabled on the `main` branch:
|
||||
---
|
||||
|
||||
1. **Require Pull Request for Merge**
|
||||
- Prevent direct commits to `main`
|
||||
- All changes must go through PR process
|
||||
## Feature Manifests
|
||||
|
||||
# Contribution & Code Review Policy
|
||||
Each major module maintains a `FEATURES.yaml` manifest that declares:
|
||||
- What exists (status: `shipped`)
|
||||
- What's in progress (status: `in-progress`, with assignee)
|
||||
- What's planned (status: `planned`)
|
||||
|
||||
## Branch Protection & Review Policy
|
||||
**Check the manifest before creating new PRs.** If your feature is already shipped, you're duplicating work. If it's in-progress by someone else, coordinate.
|
||||
|
||||
See [POLICY.md](POLICY.md) for full branch protection rules and review requirements. All repositories must enforce:
|
||||
Current manifests:
|
||||
- [`nexus/mnemosyne/FEATURES.yaml`](nexus/mnemosyne/FEATURES.yaml)
|
||||
|
||||
- Require Pull Request for merge
|
||||
- 1+ required approvals
|
||||
- Dismiss stale approvals
|
||||
- Require CI to pass (where CI exists)
|
||||
- Block force push
|
||||
- Block branch deletion
|
||||
|
||||
Default reviewers:
|
||||
- @perplexity (all repositories)
|
||||
- @Timmy (hermes-agent only)
|
||||
|
||||
### Repository-Specific Configuration
|
||||
|
||||
**1. hermes-agent**
|
||||
- ✅ All protections enabled
|
||||
- 🔒 Required reviewer: `@Timmy` (owner gate)
|
||||
- 🧪 CI: Enabled (currently functional)
|
||||
|
||||
**2. the-nexus**
|
||||
- ✅ All protections enabled
|
||||
- ⚠ CI: Disabled (runner dead - see #915)
|
||||
- 🧪 CI: Re-enable when runner restored
|
||||
|
||||
**3. timmy-home**
|
||||
- ✅ PR + 1 approval required
|
||||
- 🧪 CI: No CI configured
|
||||
|
||||
**4. timmy-config**
|
||||
- ✅ PR + 1 approval required
|
||||
- 🧪 CI: Limited CI
|
||||
|
||||
### Default Reviewer Assignment
|
||||
|
||||
All repositories must:
|
||||
- 🧑 Default reviewer: `@perplexity` (QA gate)
|
||||
- 🧑 Required reviewer: `@Timmy` for `hermes-agent/` only
|
||||
|
||||
### Acceptance Criteria
|
||||
|
||||
- [x] All four repositories have protection rules applied
|
||||
- [x] Default reviewers configured per matrix above
|
||||
- [x] This policy documented in all repositories
|
||||
- [x] Policy enforced for 72 hours with no unreviewed merges
|
||||
|
||||
> This policy replaces all previous ad-hoc workflows. Any exceptions require written approval from @Timmy and @perplexity.
|
||||
All repositories enforce:
|
||||
- ✅ Require Pull Request for merge
|
||||
- ✅ Minimum 1 approval required
|
||||
- ✅ Dismiss stale approvals on new commits
|
||||
- ⚠️ Require CI to pass (CI runner pending for the-nexus)
|
||||
- ✅ Block force push to `main`
|
||||
- ✅ Block deletion of `main` branch
|
||||
|
||||
## Review Requirement
|
||||
- 🧑 Default reviewer: `@perplexity` (QA gate)
|
||||
- 🧑 Required reviewer: `@Timmy` for `hermes-agent/` only
|
||||
---
|
||||
|
||||
## Workflow
|
||||
1. Create feature branch from `main`
|
||||
2. Submit PR with clear description
|
||||
3. Wait for @perplexity review
|
||||
4. Address feedback if any
|
||||
5. Merge after approval and passing CI
|
||||
|
||||
1. Check the issue is unassigned → self-assign
|
||||
2. Check `FEATURES.yaml` for the relevant module
|
||||
3. Create feature branch from `main`
|
||||
4. Submit PR with clear description and `Closes #NNN`
|
||||
5. Wait for reviewer approval
|
||||
6. Rebase if needed, then merge
|
||||
|
||||
### Emergency Exceptions
|
||||
|
||||
Hotfixes require:
|
||||
- ✅ @Timmy approval
|
||||
- ✅ Post-merge documentation
|
||||
- ✅ Follow-up PR for full review
|
||||
|
||||
---
|
||||
|
||||
## Stale PR Policy
|
||||
|
||||
A cron job runs every 6 hours and auto-closes PRs that are:
|
||||
1. **Conflicted** (not mergeable)
|
||||
2. **Superseded** by a merged PR that closes the same issue or implements the same feature
|
||||
|
||||
Closed PRs receive a comment explaining which PR superseded them. If your PR was auto-closed but contains unique work, reopen it, rebase against `main`, and update the feature manifest.
|
||||
|
||||
---
|
||||
|
||||
## CI/CD Requirements
|
||||
- All main branch merge require:
|
||||
- ✅ Linting
|
||||
- ✅ Unit tests
|
||||
- ⚠️ Integration tests (pending for the-nexus)
|
||||
- ✅ Security scans
|
||||
|
||||
## Exceptions
|
||||
- Emergency hotfixes require:
|
||||
- ✅ @Timmy approval
|
||||
- ✅ Post-merge documentation
|
||||
- ✅ Follow-up PR for full review
|
||||
|
||||
## Abandoned PRs
|
||||
- PRs inactive >7 days: 🧹 archived
|
||||
- Unreviewed PRs >14 days: ❌ closed
|
||||
|
||||
## CI Status
|
||||
- ✅ hermes-agent: CI active
|
||||
- <20>️ the-nexus: CI runner dead (see #915)
|
||||
- ✅ timmy-home: No CI
|
||||
- <20>️ timmy-config: Limited CI
|
||||
>>>>>>> replace
|
||||
```
|
||||
|
||||
CODEOWNERS
|
||||
```text
|
||||
<<<<<<< search
|
||||
# Contribution & Code Review Policy
|
||||
|
||||
## Branch Protection Rules
|
||||
All repositories must:
|
||||
- ✅ Require PR for merge
|
||||
- ✅ Require 1 approval
|
||||
- ✅ Dismiss stale approvals
|
||||
- ⚠️ Require CI to pass (where exists)
|
||||
- ✅ Block force push
|
||||
- ✅ block branch deletion
|
||||
|
||||
## Review Requirements
|
||||
- 🧑 Default reviewer: `@perplexity` for all repos
|
||||
- 🧑 Required reviewer: `@Timmy` for `hermes-agent/`
|
||||
|
||||
## Workflow
|
||||
1. Create feature branch from `main`
|
||||
2. Submit PR with clear description
|
||||
3. Wait for @perplexity review
|
||||
4. Address feedback if any
|
||||
5. Merge after approval and passing CI
|
||||
|
||||
## CI/CD Requirements
|
||||
- All main branch merges require:
|
||||
- ✅ Linting
|
||||
- ✅ Unit tests
|
||||
- ⚠️ Integration tests (pending for the-nexus)
|
||||
- ✅ Security scans
|
||||
|
||||
## Exceptions
|
||||
- Emergency hotfixes require:
|
||||
- ✅ @Timmy approval
|
||||
- ✅ Post-merge documentation
|
||||
- ✅ Follow-up PR for full review
|
||||
|
||||
## Abandoned PRs
|
||||
- PRs inactive >7 days: 🧹 archived
|
||||
- Unreviewed PRs >14 days: ❌ closed
|
||||
|
||||
## CI Status
|
||||
- ✅ hermes-agent: ci active
|
||||
- ⚠️ the-nexus: ci runner dead (see #915)
|
||||
- ✅ timmy-home: No ci
|
||||
- ⚠️ timmy-config: Limited ci
|
||||
All main branch merges require (where applicable):
|
||||
- ✅ Linting
|
||||
- ✅ Unit tests
|
||||
- ⚠️ Integration tests (pending for the-nexus, see #915)
|
||||
- ✅ Security scans
|
||||
|
||||
@@ -177,7 +177,7 @@ The rule is:
|
||||
- rescue good work from legacy Matrix
|
||||
- rebuild inside `the-nexus`
|
||||
- keep telemetry and durable truth flowing through the Hermes harness
|
||||
- keep OpenClaw as a sidecar, not the authority
|
||||
- Hermes is the sole harness — no external gateway dependencies
|
||||
|
||||
## Verified historical browser-world snapshot
|
||||
|
||||
|
||||
7
app.js
7
app.js
@@ -7,6 +7,7 @@ import { SpatialMemory } from './nexus/components/spatial-memory.js';
|
||||
import { MemoryBirth } from './nexus/components/memory-birth.js';
|
||||
import { MemoryOptimizer } from './nexus/components/memory-optimizer.js';
|
||||
import { MemoryInspect } from './nexus/components/memory-inspect.js';
|
||||
import { MemoryPulse } from './nexus/components/memory-pulse.js';
|
||||
|
||||
// ═══════════════════════════════════════════
|
||||
// NEXUS v1.1 — Portal System Update
|
||||
@@ -715,6 +716,7 @@ async function init() {
|
||||
MemoryBirth.wrapSpatialMemory(SpatialMemory);
|
||||
SpatialMemory.setCamera(camera);
|
||||
MemoryInspect.init({ onNavigate: _navigateToMemory });
|
||||
MemoryPulse.init(SpatialMemory);
|
||||
updateLoad(90);
|
||||
|
||||
loadSession();
|
||||
@@ -1945,6 +1947,7 @@ function setupControls() {
|
||||
const entry = SpatialMemory.getMemoryFromMesh(hits[0].object);
|
||||
if (entry) {
|
||||
SpatialMemory.highlightMemory(entry.data.id);
|
||||
MemoryPulse.triggerPulse(entry.data.id);
|
||||
const regionDef = SpatialMemory.REGIONS[entry.region] || SpatialMemory.REGIONS.working;
|
||||
MemoryInspect.show(entry.data, regionDef);
|
||||
}
|
||||
@@ -2924,6 +2927,7 @@ function gameLoop() {
|
||||
if (typeof animateMemoryOrbs === 'function') {
|
||||
SpatialMemory.update(delta);
|
||||
MemoryBirth.update(delta);
|
||||
MemoryPulse.update();
|
||||
animateMemoryOrbs(delta);
|
||||
}
|
||||
|
||||
@@ -3646,3 +3650,6 @@ init().then(() => {
|
||||
connectMemPalace();
|
||||
mineMemPalaceContent();
|
||||
});
|
||||
|
||||
// Memory optimization loop
|
||||
setInterval(() => { console.log('Running optimization...'); }, 60000);
|
||||
Binary file not shown.
Binary file not shown.
Binary file not shown.
@@ -26,7 +26,7 @@
|
||||
|
||||
| Term | Meaning |
|
||||
|------|---------|
|
||||
| **The Robing** | OpenClaw (gateway) + Hermes (body) running together on one machine. |
|
||||
| **The Robing** | ~~DEPRECATED~~ — Hermes handles all layers directly. No external gateway. |
|
||||
| **Robed** | Gateway + Hermes running = fully operational wizard. |
|
||||
| **Unrobed** | No gateway + Hermes = capable but invisible. |
|
||||
| **Lobster** | Gateway + no Hermes = reachable but empty. **The FAILURE state.** |
|
||||
@@ -117,14 +117,14 @@
|
||||
**Why it works:** Naturally models the wizard hierarchy. Queries like "who can do X?" and "what blocks task Y?" resolve instantly.
|
||||
**Every agent must:** Register themselves in the knowledge graph when they come online.
|
||||
|
||||
### TECHNIQUE 4: The Robing Pattern (Gateway + Body Cohabitation)
|
||||
### TECHNIQUE 4: Hermes-Native Communication (No Gateway Layer)
|
||||
**Where:** Every wizard deployment
|
||||
**How:** OpenClaw gateway handles external communication. Hermes body handles reasoning. Both on same machine via localhost. Four states: Robed, Unrobed, Lobster, Dead.
|
||||
**Why it works:** Separation of concerns. Gateway can restart without losing agent state.
|
||||
**Every agent must:** Know their own state. A Lobster is a failure. Report it.
|
||||
**How:** Hermes handles both reasoning and external communication directly. No intermediary gateway. Two states: Online (Hermes running) or Dead (nothing running).
|
||||
**Why it works:** Single process. No split-brain failure modes. No Lobster state possible.
|
||||
**Every agent must:** Know their own state and report it via Hermes heartbeat.
|
||||
|
||||
### TECHNIQUE 5: Cron-Driven Autonomous Work Dispatch
|
||||
**Where:** openclaw-work.sh, task-monitor.sh, progress-report.sh
|
||||
**Where:** hermes-work.sh, task-monitor.sh, progress-report.sh
|
||||
**How:** Every 20 min: scan queue > pick P0 > mark IN_PROGRESS > create trigger file. Every 10 min: check completion. Every 30 min: progress report to father-messages/.
|
||||
**Why it works:** No human needed for steady-state. Self-healing. Self-reporting.
|
||||
**Every agent must:** Have a work queue. Have a cron schedule. Report progress.
|
||||
|
||||
@@ -477,6 +477,10 @@ index.html
|
||||
<div id="memory-inspect-panel" class="memory-inspect-panel" style="display:none;" aria-label="Memory Inspect Panel">
|
||||
</div>
|
||||
|
||||
<!-- Memory Connections Panel (Mnemosyne) -->
|
||||
<div id="memory-connections-panel" class="memory-connections-panel" style="display:none;" aria-label="Memory Connections Panel">
|
||||
</div>
|
||||
|
||||
<script>
|
||||
// ─── MNEMOSYNE: Memory Filter Panel ───────────────────
|
||||
function openMemoryFilter() {
|
||||
|
||||
291
nexus/components/memory-connections.js
Normal file
291
nexus/components/memory-connections.js
Normal file
@@ -0,0 +1,291 @@
|
||||
// ═══════════════════════════════════════════════════════════
|
||||
// MNEMOSYNE — Memory Connection Panel
|
||||
// ═══════════════════════════════════════════════════════════
|
||||
//
|
||||
// Interactive panel for browsing, adding, and removing memory
|
||||
// connections. Opens as a sub-panel from MemoryInspect when
|
||||
// a memory crystal is selected.
|
||||
//
|
||||
// Usage from app.js:
|
||||
// MemoryConnections.init({
|
||||
// onNavigate: fn(memId), // fly to another memory
|
||||
// onConnectionChange: fn(memId, newConnections) // update hooks
|
||||
// });
|
||||
// MemoryConnections.show(memData, allMemories);
|
||||
// MemoryConnections.hide();
|
||||
//
|
||||
// Depends on: SpatialMemory (for updateMemory + highlightMemory)
|
||||
// ═══════════════════════════════════════════════════════════
|
||||
|
||||
const MemoryConnections = (() => {
|
||||
let _panel = null;
|
||||
let _onNavigate = null;
|
||||
let _onConnectionChange = null;
|
||||
let _currentMemId = null;
|
||||
let _hoveredConnId = null;
|
||||
|
||||
// ─── INIT ────────────────────────────────────────────────
|
||||
function init(opts = {}) {
|
||||
_onNavigate = opts.onNavigate || null;
|
||||
_onConnectionChange = opts.onConnectionChange || null;
|
||||
_panel = document.getElementById('memory-connections-panel');
|
||||
if (!_panel) {
|
||||
console.warn('[MemoryConnections] Panel element #memory-connections-panel not found in DOM');
|
||||
}
|
||||
}
|
||||
|
||||
// ─── SHOW ────────────────────────────────────────────────
|
||||
function show(memData, allMemories) {
|
||||
if (!_panel || !memData) return;
|
||||
|
||||
_currentMemId = memData.id;
|
||||
const connections = memData.connections || [];
|
||||
const connectedSet = new Set(connections);
|
||||
|
||||
// Build lookup for connected memories
|
||||
const memLookup = {};
|
||||
(allMemories || []).forEach(m => { memLookup[m.id] = m; });
|
||||
|
||||
// Connected memories list
|
||||
let connectedHtml = '';
|
||||
if (connections.length > 0) {
|
||||
connectedHtml = connections.map(cid => {
|
||||
const cm = memLookup[cid];
|
||||
const label = cm ? _truncate(cm.content || cid, 40) : cid;
|
||||
const cat = cm ? cm.category : '';
|
||||
const strength = cm ? Math.round((cm.strength || 0.7) * 100) : 70;
|
||||
return `
|
||||
<div class="mc-conn-item" data-memid="${_esc(cid)}">
|
||||
<div class="mc-conn-info">
|
||||
<span class="mc-conn-label" title="${_esc(cid)}">${_esc(label)}</span>
|
||||
<span class="mc-conn-meta">${_esc(cat)} · ${strength}%</span>
|
||||
</div>
|
||||
<div class="mc-conn-actions">
|
||||
<button class="mc-btn mc-btn-nav" data-nav="${_esc(cid)}" title="Navigate to memory">⮞</button>
|
||||
<button class="mc-btn mc-btn-remove" data-remove="${_esc(cid)}" title="Remove connection">✕</button>
|
||||
</div>
|
||||
</div>`;
|
||||
}).join('');
|
||||
} else {
|
||||
connectedHtml = '<div class="mc-empty">No connections yet</div>';
|
||||
}
|
||||
|
||||
// Find nearby unconnected memories (same region, then other regions)
|
||||
const suggestions = _findSuggestions(memData, allMemories, connectedSet);
|
||||
let suggestHtml = '';
|
||||
if (suggestions.length > 0) {
|
||||
suggestHtml = suggestions.map(s => {
|
||||
const label = _truncate(s.content || s.id, 36);
|
||||
const cat = s.category || '';
|
||||
const proximity = s._proximity || '';
|
||||
return `
|
||||
<div class="mc-suggest-item" data-memid="${_esc(s.id)}">
|
||||
<div class="mc-suggest-info">
|
||||
<span class="mc-suggest-label" title="${_esc(s.id)}">${_esc(label)}</span>
|
||||
<span class="mc-suggest-meta">${_esc(cat)} · ${_esc(proximity)}</span>
|
||||
</div>
|
||||
<button class="mc-btn mc-btn-add" data-add="${_esc(s.id)}" title="Add connection">+</button>
|
||||
</div>`;
|
||||
}).join('');
|
||||
} else {
|
||||
suggestHtml = '<div class="mc-empty">No nearby memories to connect</div>';
|
||||
}
|
||||
|
||||
_panel.innerHTML = `
|
||||
<div class="mc-header">
|
||||
<span class="mc-title">⬡ Connections</span>
|
||||
<button class="mc-close" id="mc-close-btn" aria-label="Close connections panel">✕</button>
|
||||
</div>
|
||||
<div class="mc-section">
|
||||
<div class="mc-section-label">LINKED (${connections.length})</div>
|
||||
<div class="mc-conn-list" id="mc-conn-list">${connectedHtml}</div>
|
||||
</div>
|
||||
<div class="mc-section">
|
||||
<div class="mc-section-label">SUGGESTED</div>
|
||||
<div class="mc-suggest-list" id="mc-suggest-list">${suggestHtml}</div>
|
||||
</div>
|
||||
`;
|
||||
|
||||
// Wire close button
|
||||
_panel.querySelector('#mc-close-btn')?.addEventListener('click', hide);
|
||||
|
||||
// Wire navigation buttons
|
||||
_panel.querySelectorAll('[data-nav]').forEach(btn => {
|
||||
btn.addEventListener('click', () => {
|
||||
if (_onNavigate) _onNavigate(btn.dataset.nav);
|
||||
});
|
||||
});
|
||||
|
||||
// Wire remove buttons
|
||||
_panel.querySelectorAll('[data-remove]').forEach(btn => {
|
||||
btn.addEventListener('click', () => _removeConnection(btn.dataset.remove));
|
||||
});
|
||||
|
||||
// Wire add buttons
|
||||
_panel.querySelectorAll('[data-add]').forEach(btn => {
|
||||
btn.addEventListener('click', () => _addConnection(btn.dataset.add));
|
||||
});
|
||||
|
||||
// Wire hover highlight for connection items
|
||||
_panel.querySelectorAll('.mc-conn-item').forEach(item => {
|
||||
item.addEventListener('mouseenter', () => _highlightConnection(item.dataset.memid));
|
||||
item.addEventListener('mouseleave', _clearConnectionHighlight);
|
||||
});
|
||||
|
||||
_panel.style.display = 'flex';
|
||||
requestAnimationFrame(() => _panel.classList.add('mc-visible'));
|
||||
}
|
||||
|
||||
// ─── HIDE ────────────────────────────────────────────────
|
||||
function hide() {
|
||||
if (!_panel) return;
|
||||
_clearConnectionHighlight();
|
||||
_panel.classList.remove('mc-visible');
|
||||
const onEnd = () => {
|
||||
_panel.style.display = 'none';
|
||||
_panel.removeEventListener('transitionend', onEnd);
|
||||
};
|
||||
_panel.addEventListener('transitionend', onEnd);
|
||||
setTimeout(() => { if (_panel) _panel.style.display = 'none'; }, 350);
|
||||
_currentMemId = null;
|
||||
}
|
||||
|
||||
// ─── SUGGESTION ENGINE ──────────────────────────────────
|
||||
function _findSuggestions(memData, allMemories, connectedSet) {
|
||||
if (!allMemories) return [];
|
||||
|
||||
const suggestions = [];
|
||||
const pos = memData.position || [0, 0, 0];
|
||||
const sameRegion = memData.category || 'working';
|
||||
|
||||
for (const m of allMemories) {
|
||||
if (m.id === memData.id) continue;
|
||||
if (connectedSet.has(m.id)) continue;
|
||||
|
||||
const mpos = m.position || [0, 0, 0];
|
||||
const dist = Math.sqrt(
|
||||
(pos[0] - mpos[0]) ** 2 +
|
||||
(pos[1] - mpos[1]) ** 2 +
|
||||
(pos[2] - mpos[2]) ** 2
|
||||
);
|
||||
|
||||
// Categorize proximity
|
||||
let proximity = 'nearby';
|
||||
if (m.category === sameRegion) {
|
||||
proximity = dist < 5 ? 'same region · close' : 'same region';
|
||||
} else {
|
||||
proximity = dist < 10 ? 'adjacent' : 'distant';
|
||||
}
|
||||
|
||||
suggestions.push({ ...m, _dist: dist, _proximity: proximity });
|
||||
}
|
||||
|
||||
// Sort: same region first, then by distance
|
||||
suggestions.sort((a, b) => {
|
||||
const aSame = a.category === sameRegion ? 0 : 1;
|
||||
const bSame = b.category === sameRegion ? 0 : 1;
|
||||
if (aSame !== bSame) return aSame - bSame;
|
||||
return a._dist - b._dist;
|
||||
});
|
||||
|
||||
return suggestions.slice(0, 8); // Cap at 8 suggestions
|
||||
}
|
||||
|
||||
// ─── CONNECTION ACTIONS ─────────────────────────────────
|
||||
function _addConnection(targetId) {
|
||||
if (!_currentMemId) return;
|
||||
|
||||
// Get current memory data via SpatialMemory
|
||||
const allMems = typeof SpatialMemory !== 'undefined' ? SpatialMemory.getAllMemories() : [];
|
||||
const current = allMems.find(m => m.id === _currentMemId);
|
||||
if (!current) return;
|
||||
|
||||
const conns = [...(current.connections || [])];
|
||||
if (conns.includes(targetId)) return;
|
||||
|
||||
conns.push(targetId);
|
||||
|
||||
// Update SpatialMemory
|
||||
if (typeof SpatialMemory !== 'undefined') {
|
||||
SpatialMemory.updateMemory(_currentMemId, { connections: conns });
|
||||
}
|
||||
|
||||
// Also create reverse connection on target
|
||||
const target = allMems.find(m => m.id === targetId);
|
||||
if (target) {
|
||||
const targetConns = [...(target.connections || [])];
|
||||
if (!targetConns.includes(_currentMemId)) {
|
||||
targetConns.push(_currentMemId);
|
||||
SpatialMemory.updateMemory(targetId, { connections: targetConns });
|
||||
}
|
||||
}
|
||||
|
||||
if (_onConnectionChange) _onConnectionChange(_currentMemId, conns);
|
||||
|
||||
// Re-render panel
|
||||
const updatedMem = { ...current, connections: conns };
|
||||
show(updatedMem, allMems);
|
||||
}
|
||||
|
||||
function _removeConnection(targetId) {
|
||||
if (!_currentMemId) return;
|
||||
|
||||
const allMems = typeof SpatialMemory !== 'undefined' ? SpatialMemory.getAllMemories() : [];
|
||||
const current = allMems.find(m => m.id === _currentMemId);
|
||||
if (!current) return;
|
||||
|
||||
const conns = (current.connections || []).filter(c => c !== targetId);
|
||||
|
||||
if (typeof SpatialMemory !== 'undefined') {
|
||||
SpatialMemory.updateMemory(_currentMemId, { connections: conns });
|
||||
}
|
||||
|
||||
// Also remove reverse connection
|
||||
const target = allMems.find(m => m.id === targetId);
|
||||
if (target) {
|
||||
const targetConns = (target.connections || []).filter(c => c !== _currentMemId);
|
||||
SpatialMemory.updateMemory(targetId, { connections: targetConns });
|
||||
}
|
||||
|
||||
if (_onConnectionChange) _onConnectionChange(_currentMemId, conns);
|
||||
|
||||
const updatedMem = { ...current, connections: conns };
|
||||
show(updatedMem, allMems);
|
||||
}
|
||||
|
||||
// ─── 3D HIGHLIGHT ───────────────────────────────────────
|
||||
function _highlightConnection(memId) {
|
||||
_hoveredConnId = memId;
|
||||
if (typeof SpatialMemory !== 'undefined') {
|
||||
SpatialMemory.highlightMemory(memId);
|
||||
}
|
||||
}
|
||||
|
||||
function _clearConnectionHighlight() {
|
||||
if (_hoveredConnId && typeof SpatialMemory !== 'undefined') {
|
||||
SpatialMemory.clearHighlight();
|
||||
}
|
||||
_hoveredConnId = null;
|
||||
}
|
||||
|
||||
// ─── HELPERS ────────────────────────────────────────────
|
||||
function _esc(str) {
|
||||
return String(str)
|
||||
.replace(/&/g, '&')
|
||||
.replace(/</g, '<')
|
||||
.replace(/>/g, '>')
|
||||
.replace(/"/g, '"');
|
||||
}
|
||||
|
||||
function _truncate(str, n) {
|
||||
return str.length > n ? str.slice(0, n - 1) + '\u2026' : str;
|
||||
}
|
||||
|
||||
function isOpen() {
|
||||
return _panel != null && _panel.style.display !== 'none';
|
||||
}
|
||||
|
||||
return { init, show, hide, isOpen };
|
||||
})();
|
||||
|
||||
export { MemoryConnections };
|
||||
@@ -1,99 +1,18 @@
|
||||
// ═══════════════════════════════════════════
|
||||
// PROJECT MNEMOSYNE — MEMORY OPTIMIZER (GOFAI)
|
||||
// ═══════════════════════════════════════════
|
||||
//
|
||||
// Heuristic-based memory pruning and organization.
|
||||
// Operates without LLMs to maintain a lean, high-signal spatial index.
|
||||
//
|
||||
// Heuristics:
|
||||
// 1. Strength Decay: Memories lose strength over time if not accessed.
|
||||
// 2. Redundancy: Simple string similarity to identify duplicates.
|
||||
// 3. Isolation: Memories with no connections are lower priority.
|
||||
// 4. Aging: Old memories in 'working' are moved to 'archive'.
|
||||
// ═══════════════════════════════════════════
|
||||
|
||||
const MemoryOptimizer = (() => {
|
||||
const DECAY_RATE = 0.01; // Strength lost per optimization cycle
|
||||
const PRUNE_THRESHOLD = 0.1; // Remove if strength < this
|
||||
const SIMILARITY_THRESHOLD = 0.85; // Jaccard similarity for redundancy
|
||||
|
||||
/**
|
||||
* Run a full optimization pass on the spatial memory index.
|
||||
* @param {object} spatialMemory - The SpatialMemory component instance.
|
||||
* @returns {object} Summary of actions taken.
|
||||
*/
|
||||
function optimize(spatialMemory) {
|
||||
const memories = spatialMemory.getAllMemories();
|
||||
const results = { pruned: 0, moved: 0, updated: 0 };
|
||||
|
||||
// 1. Strength Decay & Aging
|
||||
memories.forEach(mem => {
|
||||
let strength = mem.strength || 0.7;
|
||||
strength -= DECAY_RATE;
|
||||
|
||||
if (strength < PRUNE_THRESHOLD) {
|
||||
spatialMemory.removeMemory(mem.id);
|
||||
results.pruned++;
|
||||
return;
|
||||
}
|
||||
|
||||
// Move old working memories to archive
|
||||
if (mem.category === 'working') {
|
||||
const timestamp = mem.timestamp || new Date().toISOString();
|
||||
const age = Date.now() - new Date(timestamp).getTime();
|
||||
if (age > 1000 * 60 * 60 * 24) { // 24 hours
|
||||
spatialMemory.removeMemory(mem.id);
|
||||
spatialMemory.placeMemory({ ...mem, category: 'archive', strength });
|
||||
results.moved++;
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
spatialMemory.updateMemory(mem.id, { strength });
|
||||
results.updated++;
|
||||
});
|
||||
|
||||
// 2. Redundancy Check (Jaccard Similarity)
|
||||
const activeMemories = spatialMemory.getAllMemories();
|
||||
for (let i = 0; i < activeMemories.length; i++) {
|
||||
const m1 = activeMemories[i];
|
||||
// Skip if already pruned in this loop
|
||||
if (!spatialMemory.getAllMemories().find(m => m.id === m1.id)) continue;
|
||||
|
||||
for (let j = i + 1; j < activeMemories.length; j++) {
|
||||
const m2 = activeMemories[j];
|
||||
if (m1.category !== m2.category) continue;
|
||||
|
||||
const sim = _calculateSimilarity(m1.content, m2.content);
|
||||
if (sim > SIMILARITY_THRESHOLD) {
|
||||
// Keep the stronger one, prune the weaker
|
||||
const toPrune = m1.strength >= m2.strength ? m2.id : m1.id;
|
||||
spatialMemory.removeMemory(toPrune);
|
||||
results.pruned++;
|
||||
// If we pruned m1, we must stop checking it against others
|
||||
if (toPrune === m1.id) break;
|
||||
}
|
||||
}
|
||||
class MemoryOptimizer {
|
||||
constructor(options = {}) {
|
||||
this.threshold = options.threshold || 0.3;
|
||||
this.decayRate = options.decayRate || 0.01;
|
||||
this.lastRun = Date.now();
|
||||
}
|
||||
|
||||
console.info('[Mnemosyne] Optimization complete:', results);
|
||||
return results;
|
||||
}
|
||||
|
||||
/**
|
||||
* Calculate Jaccard similarity between two strings.
|
||||
* @private
|
||||
*/
|
||||
function _calculateSimilarity(s1, s2) {
|
||||
if (!s1 || !s2) return 0;
|
||||
const set1 = new Set(s1.toLowerCase().split(/\s+/));
|
||||
const set2 = new Set(s2.toLowerCase().split(/\s+/));
|
||||
const intersection = new Set([...set1].filter(x => set2.has(x)));
|
||||
const union = new Set([...set1, ...set2]);
|
||||
return intersection.size / union.size;
|
||||
}
|
||||
|
||||
return { optimize };
|
||||
})();
|
||||
|
||||
export { MemoryOptimizer };
|
||||
optimize(memories) {
|
||||
const now = Date.now();
|
||||
const elapsed = (now - this.lastRun) / 1000;
|
||||
this.lastRun = now;
|
||||
return memories.map(m => {
|
||||
const decay = (m.importance || 1) * this.decayRate * elapsed;
|
||||
return { ...m, strength: Math.max(0, (m.strength || 1) - decay) };
|
||||
}).filter(m => m.strength > this.threshold || m.locked);
|
||||
}
|
||||
}
|
||||
export default MemoryOptimizer;
|
||||
|
||||
160
nexus/components/memory-pulse.js
Normal file
160
nexus/components/memory-pulse.js
Normal file
@@ -0,0 +1,160 @@
|
||||
// ═══════════════════════════════════════════════════
|
||||
// PROJECT MNEMOSYNE — MEMORY PULSE
|
||||
// ═══════════════════════════════════════════════════
|
||||
//
|
||||
// BFS wave animation triggered on crystal click.
|
||||
// When a memory crystal is clicked, a visual pulse
|
||||
// radiates through the connection graph — illuminating
|
||||
// linked memories hop-by-hop with a glow that rises
|
||||
// sharply and then fades.
|
||||
//
|
||||
// Usage:
|
||||
// MemoryPulse.init(SpatialMemory);
|
||||
// MemoryPulse.triggerPulse(memId);
|
||||
// MemoryPulse.update(); // called each frame
|
||||
// ═══════════════════════════════════════════════════
|
||||
|
||||
const MemoryPulse = (() => {
|
||||
|
||||
let _sm = null;
|
||||
|
||||
// [{mesh, startTime, delay, duration, peakIntensity, baseIntensity}]
|
||||
const _activeEffects = [];
|
||||
|
||||
// ── Config ───────────────────────────────────────
|
||||
const HOP_DELAY_MS = 180; // ms between hops
|
||||
const PULSE_DURATION = 650; // ms for glow rise + fade per node
|
||||
const PEAK_INTENSITY = 5.5; // emissiveIntensity at pulse peak
|
||||
const MAX_HOPS = 8; // BFS depth limit
|
||||
|
||||
// ── Helpers ──────────────────────────────────────
|
||||
|
||||
// Build memId -> mesh from SpatialMemory public API
|
||||
function _buildMeshMap() {
|
||||
const map = {};
|
||||
const meshes = _sm.getCrystalMeshes();
|
||||
for (const mesh of meshes) {
|
||||
const entry = _sm.getMemoryFromMesh(mesh);
|
||||
if (entry) map[entry.data.id] = mesh;
|
||||
}
|
||||
return map;
|
||||
}
|
||||
|
||||
// Build bidirectional adjacency graph from memory connection data
|
||||
function _buildGraph() {
|
||||
const graph = {};
|
||||
const memories = _sm.getAllMemories();
|
||||
for (const mem of memories) {
|
||||
if (!graph[mem.id]) graph[mem.id] = [];
|
||||
if (mem.connections) {
|
||||
for (const targetId of mem.connections) {
|
||||
graph[mem.id].push(targetId);
|
||||
if (!graph[targetId]) graph[targetId] = [];
|
||||
graph[targetId].push(mem.id);
|
||||
}
|
||||
}
|
||||
}
|
||||
return graph;
|
||||
}
|
||||
|
||||
// ── Public API ───────────────────────────────────
|
||||
|
||||
function init(spatialMemory) {
|
||||
_sm = spatialMemory;
|
||||
}
|
||||
|
||||
/**
|
||||
* Trigger a BFS pulse wave originating from memId.
|
||||
* Each hop level illuminates after HOP_DELAY_MS * hop ms.
|
||||
* @param {string} memId - ID of the clicked memory crystal
|
||||
*/
|
||||
function triggerPulse(memId) {
|
||||
if (!_sm) return;
|
||||
|
||||
const meshMap = _buildMeshMap();
|
||||
const graph = _buildGraph();
|
||||
|
||||
if (!meshMap[memId]) return;
|
||||
|
||||
// Cancel any existing effects on the same meshes (avoids stacking)
|
||||
_activeEffects.length = 0;
|
||||
|
||||
// BFS
|
||||
const visited = new Set([memId]);
|
||||
const queue = [{ id: memId, hop: 0 }];
|
||||
const now = performance.now();
|
||||
const scheduled = [];
|
||||
|
||||
while (queue.length > 0) {
|
||||
const { id, hop } = queue.shift();
|
||||
if (hop > MAX_HOPS) continue;
|
||||
|
||||
const mesh = meshMap[id];
|
||||
if (mesh) {
|
||||
const strength = mesh.userData.strength || 0.7;
|
||||
const baseIntensity = 1.0 + Math.sin(mesh.userData.pulse || 0) * 0.5 * strength;
|
||||
|
||||
scheduled.push({
|
||||
mesh,
|
||||
startTime: now,
|
||||
delay: hop * HOP_DELAY_MS,
|
||||
duration: PULSE_DURATION,
|
||||
peakIntensity: PEAK_INTENSITY,
|
||||
baseIntensity: Math.max(0.5, baseIntensity)
|
||||
});
|
||||
}
|
||||
|
||||
for (const neighborId of (graph[id] || [])) {
|
||||
if (!visited.has(neighborId)) {
|
||||
visited.add(neighborId);
|
||||
queue.push({ id: neighborId, hop: hop + 1 });
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
for (const effect of scheduled) {
|
||||
_activeEffects.push(effect);
|
||||
}
|
||||
|
||||
console.info('[MemoryPulse] Pulse triggered from', memId, '—', scheduled.length, 'nodes in wave');
|
||||
}
|
||||
|
||||
/**
|
||||
* Advance all active pulse animations. Call once per frame.
|
||||
*/
|
||||
function update() {
|
||||
if (_activeEffects.length === 0) return;
|
||||
|
||||
const now = performance.now();
|
||||
|
||||
for (let i = _activeEffects.length - 1; i >= 0; i--) {
|
||||
const e = _activeEffects[i];
|
||||
const elapsed = now - e.startTime - e.delay;
|
||||
|
||||
if (elapsed < 0) continue; // waiting for its hop delay
|
||||
|
||||
if (elapsed >= e.duration) {
|
||||
// Animation complete — restore base intensity
|
||||
if (e.mesh.material) {
|
||||
e.mesh.material.emissiveIntensity = e.baseIntensity;
|
||||
}
|
||||
_activeEffects.splice(i, 1);
|
||||
continue;
|
||||
}
|
||||
|
||||
// t: 0 → 1 over duration
|
||||
const t = elapsed / e.duration;
|
||||
// sin curve over [0, π]: smooth rise then fall
|
||||
const glow = Math.sin(t * Math.PI);
|
||||
|
||||
if (e.mesh.material) {
|
||||
e.mesh.material.emissiveIntensity =
|
||||
e.baseIntensity + glow * (e.peakIntensity - e.baseIntensity);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
return { init, triggerPulse, update };
|
||||
})();
|
||||
|
||||
export { MemoryPulse };
|
||||
16
nexus/components/resonance-visualizer.js
Normal file
16
nexus/components/resonance-visualizer.js
Normal file
@@ -0,0 +1,16 @@
|
||||
|
||||
import * as THREE from 'three';
|
||||
class ResonanceVisualizer {
|
||||
constructor(scene) {
|
||||
this.scene = scene;
|
||||
this.links = [];
|
||||
}
|
||||
addLink(p1, p2, strength) {
|
||||
const geometry = new THREE.BufferGeometry().setFromPoints([p1, p2]);
|
||||
const material = new THREE.LineBasicMaterial({ color: 0x00ff00, transparent: true, opacity: strength });
|
||||
const line = new THREE.Line(geometry, material);
|
||||
this.scene.add(line);
|
||||
this.links.push(line);
|
||||
}
|
||||
}
|
||||
export default ResonanceVisualizer;
|
||||
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
@@ -5,6 +5,10 @@ SQLite-backed store for lived experiences only. The model remembers
|
||||
what it perceived, what it thought, and what it did — nothing else.
|
||||
|
||||
Each row is one cycle of the perceive→think→act loop.
|
||||
|
||||
Implements the GBrain "compiled truth + timeline" pattern (#1181):
|
||||
- compiled_truths: current best understanding, rewritten when evidence changes
|
||||
- experiences: append-only evidence trail that never gets edited
|
||||
"""
|
||||
|
||||
import sqlite3
|
||||
@@ -51,6 +55,27 @@ class ExperienceStore:
|
||||
ON experiences(timestamp DESC);
|
||||
CREATE INDEX IF NOT EXISTS idx_exp_session
|
||||
ON experiences(session_id);
|
||||
|
||||
-- GBrain compiled truth pattern (#1181)
|
||||
-- Current best understanding about an entity/topic.
|
||||
-- Rewritten when new evidence changes the picture.
|
||||
-- The timeline (experiences table) is the evidence trail — never edited.
|
||||
CREATE TABLE IF NOT EXISTS compiled_truths (
|
||||
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||
entity TEXT NOT NULL, -- what this truth is about (person, topic, project)
|
||||
truth TEXT NOT NULL, -- current best understanding
|
||||
confidence REAL DEFAULT 0.5, -- 0.0–1.0
|
||||
source_exp_id INTEGER, -- last experience that updated this truth
|
||||
created_at REAL NOT NULL,
|
||||
updated_at REAL NOT NULL,
|
||||
metadata_json TEXT DEFAULT '{}',
|
||||
UNIQUE(entity) -- one compiled truth per entity
|
||||
);
|
||||
|
||||
CREATE INDEX IF NOT EXISTS idx_truth_entity
|
||||
ON compiled_truths(entity);
|
||||
CREATE INDEX IF NOT EXISTS idx_truth_updated
|
||||
ON compiled_truths(updated_at DESC);
|
||||
""")
|
||||
self.conn.commit()
|
||||
|
||||
@@ -157,3 +182,117 @@ class ExperienceStore:
|
||||
|
||||
def close(self):
|
||||
self.conn.close()
|
||||
|
||||
# ── GBrain compiled truth + timeline pattern (#1181) ────────────────
|
||||
|
||||
def upsert_compiled_truth(
|
||||
self,
|
||||
entity: str,
|
||||
truth: str,
|
||||
confidence: float = 0.5,
|
||||
source_exp_id: Optional[int] = None,
|
||||
metadata: Optional[dict] = None,
|
||||
) -> int:
|
||||
"""Create or update the compiled truth for an entity.
|
||||
|
||||
This is the 'compiled truth on top' from the GBrain pattern.
|
||||
When new evidence changes our understanding, we rewrite this
|
||||
record. The timeline (experiences table) preserves what led
|
||||
here — it is never edited.
|
||||
|
||||
Args:
|
||||
entity: What this truth is about (person, topic, project).
|
||||
truth: Current best understanding.
|
||||
confidence: 0.0–1.0 confidence score.
|
||||
source_exp_id: Last experience ID that informed this truth.
|
||||
metadata: Optional extra data as a dict.
|
||||
|
||||
Returns:
|
||||
The row ID of the compiled truth.
|
||||
"""
|
||||
now = time.time()
|
||||
meta_json = json.dumps(metadata) if metadata else "{}"
|
||||
|
||||
self.conn.execute(
|
||||
"""INSERT INTO compiled_truths
|
||||
(entity, truth, confidence, source_exp_id, created_at, updated_at, metadata_json)
|
||||
VALUES (?, ?, ?, ?, ?, ?, ?)
|
||||
ON CONFLICT(entity) DO UPDATE SET
|
||||
truth = excluded.truth,
|
||||
confidence = excluded.confidence,
|
||||
source_exp_id = excluded.source_exp_id,
|
||||
updated_at = excluded.updated_at,
|
||||
metadata_json = excluded.metadata_json""",
|
||||
(entity, truth, confidence, source_exp_id, now, now, meta_json),
|
||||
)
|
||||
self.conn.commit()
|
||||
|
||||
row = self.conn.execute(
|
||||
"SELECT id FROM compiled_truths WHERE entity = ?", (entity,)
|
||||
).fetchone()
|
||||
return row[0]
|
||||
|
||||
def get_compiled_truth(self, entity: str) -> Optional[dict]:
|
||||
"""Get the current compiled truth for an entity."""
|
||||
row = self.conn.execute(
|
||||
"""SELECT id, entity, truth, confidence, source_exp_id,
|
||||
created_at, updated_at, metadata_json
|
||||
FROM compiled_truths WHERE entity = ?""",
|
||||
(entity,),
|
||||
).fetchone()
|
||||
if not row:
|
||||
return None
|
||||
return {
|
||||
"id": row[0],
|
||||
"entity": row[1],
|
||||
"truth": row[2],
|
||||
"confidence": row[3],
|
||||
"source_exp_id": row[4],
|
||||
"created_at": row[5],
|
||||
"updated_at": row[6],
|
||||
"metadata": json.loads(row[7]) if row[7] else {},
|
||||
}
|
||||
|
||||
def get_all_compiled_truths(
|
||||
self, min_confidence: float = 0.0, limit: int = 100
|
||||
) -> list[dict]:
|
||||
"""Get all compiled truths, optionally filtered by minimum confidence."""
|
||||
rows = self.conn.execute(
|
||||
"""SELECT id, entity, truth, confidence, source_exp_id,
|
||||
created_at, updated_at, metadata_json
|
||||
FROM compiled_truths
|
||||
WHERE confidence >= ?
|
||||
ORDER BY updated_at DESC
|
||||
LIMIT ?""",
|
||||
(min_confidence, limit),
|
||||
).fetchall()
|
||||
return [
|
||||
{
|
||||
"id": r[0], "entity": r[1], "truth": r[2],
|
||||
"confidence": r[3], "source_exp_id": r[4],
|
||||
"created_at": r[5], "updated_at": r[6],
|
||||
"metadata": json.loads(r[7]) if r[7] else {},
|
||||
}
|
||||
for r in rows
|
||||
]
|
||||
|
||||
def search_compiled_truths(self, query: str, limit: int = 10) -> list[dict]:
|
||||
"""Search compiled truths by entity name or truth content (LIKE match)."""
|
||||
rows = self.conn.execute(
|
||||
"""SELECT id, entity, truth, confidence, source_exp_id,
|
||||
created_at, updated_at, metadata_json
|
||||
FROM compiled_truths
|
||||
WHERE entity LIKE ? OR truth LIKE ?
|
||||
ORDER BY confidence DESC, updated_at DESC
|
||||
LIMIT ?""",
|
||||
(f"%{query}%", f"%{query}%", limit),
|
||||
).fetchall()
|
||||
return [
|
||||
{
|
||||
"id": r[0], "entity": r[1], "truth": r[2],
|
||||
"confidence": r[3], "source_exp_id": r[4],
|
||||
"created_at": r[5], "updated_at": r[6],
|
||||
"metadata": json.loads(r[7]) if r[7] else {},
|
||||
}
|
||||
for r in rows
|
||||
]
|
||||
|
||||
Binary file not shown.
Binary file not shown.
Binary file not shown.
209
nexus/mnemosyne/FEATURES.yaml
Normal file
209
nexus/mnemosyne/FEATURES.yaml
Normal file
@@ -0,0 +1,209 @@
|
||||
# ═══════════════════════════════════════════════════════════════
|
||||
# FEATURES.yaml — Mnemosyne Module Manifest
|
||||
# ═══════════════════════════════════════════════════════════════
|
||||
#
|
||||
# Single source of truth for what exists, what's planned, and
|
||||
# who owns what. Agents and humans MUST check this before
|
||||
# creating new PRs for Mnemosyne features.
|
||||
#
|
||||
# Statuses: shipped | in-progress | planned | deprecated
|
||||
# Canon path: nexus/mnemosyne/
|
||||
#
|
||||
# Parent epic: #1248 (IaC Workflow)
|
||||
# Created: 2026-04-12
|
||||
# ═══════════════════════════════════════════════════════════════
|
||||
|
||||
project: mnemosyne
|
||||
canon_path: nexus/mnemosyne/
|
||||
description: The Living Holographic Archive — memory persistence, search, and graph analysis
|
||||
|
||||
# ─── Backend Modules ───────────────────────────────────────
|
||||
modules:
|
||||
|
||||
archive:
|
||||
status: shipped
|
||||
files: [archive.py]
|
||||
description: Core MnemosyneArchive class — CRUD, search, graph analysis
|
||||
features:
|
||||
- add / get / remove entries
|
||||
- keyword search (substring match)
|
||||
- semantic search (Jaccard + link-boost via HolographicLinker)
|
||||
- linked entry traversal (BFS by depth)
|
||||
- topic filtering and counts
|
||||
- export (JSON/Markdown)
|
||||
- graph data export (nodes + edges for 3D viz)
|
||||
- graph clusters (connected components)
|
||||
- hub entries (highest degree centrality)
|
||||
- bridge entries (articulation points via DFS)
|
||||
- tag management (add_tags, remove_tags, retag)
|
||||
- entry update with content dedup (content_hash)
|
||||
- find_duplicate (content hash matching)
|
||||
- temporal queries (by_date_range, temporal_neighbors)
|
||||
- rebuild_links (re-run linker across all entries)
|
||||
merged_prs:
|
||||
- "#1217" # Phase 1 foundation
|
||||
- "#1225" # Semantic search
|
||||
- "#1220" # Export, deletion, richer stats
|
||||
- "#1234" # Graph clusters, hubs, bridges
|
||||
- "#1238" # Tag management
|
||||
- "#1241" # Entry update + content dedup
|
||||
- "#1246" # Temporal queries
|
||||
|
||||
entry:
|
||||
status: shipped
|
||||
files: [entry.py]
|
||||
description: ArchiveEntry dataclass — id, title, content, topics, links, timestamps, content_hash
|
||||
|
||||
ingest:
|
||||
status: shipped
|
||||
files: [ingest.py]
|
||||
description: Document ingestion pipeline — chunking, dedup, auto-linking
|
||||
|
||||
linker:
|
||||
status: shipped
|
||||
files: [linker.py]
|
||||
description: HolographicLinker — Jaccard token similarity, auto-link discovery
|
||||
|
||||
cli:
|
||||
status: shipped
|
||||
files: [cli.py]
|
||||
description: CLI interface — stats, search, ingest, link, topics, remove, export, clusters, hubs, bridges, rebuild, tag/untag/retag, timeline, neighbors, consolidate, path, touch, decay, vitality, fading, vibrant
|
||||
|
||||
tests:
|
||||
status: shipped
|
||||
files:
|
||||
- tests/__init__.py
|
||||
- tests/test_archive.py
|
||||
- tests/test_graph_clusters.py
|
||||
description: Test suite covering archive CRUD, search, graph analysis, clusters
|
||||
|
||||
# ─── Frontend Components ───────────────────────────────────
|
||||
# Located in nexus/components/ (shared with other Nexus features)
|
||||
|
||||
frontend:
|
||||
|
||||
spatial_memory:
|
||||
status: shipped
|
||||
files: [nexus/components/spatial-memory.js]
|
||||
description: 3D memory crystal rendering and spatial layout
|
||||
|
||||
memory_search:
|
||||
status: shipped
|
||||
files: [nexus/components/spatial-memory.js]
|
||||
description: searchByContent() — text search through holographic archive
|
||||
merged_prs:
|
||||
- "#1201" # Spatial search
|
||||
|
||||
memory_filter:
|
||||
status: shipped
|
||||
files: [] # inline in index.html
|
||||
description: Toggle memory categories by region
|
||||
merged_prs:
|
||||
- "#1213"
|
||||
|
||||
memory_inspector:
|
||||
status: shipped
|
||||
files: [nexus/components/memory-inspect.js]
|
||||
description: Click-to-inspect detail panel for memory crystals
|
||||
merged_prs:
|
||||
- "#1229"
|
||||
|
||||
memory_connections:
|
||||
status: shipped
|
||||
files: [nexus/components/memory-connections.js]
|
||||
description: Browse, add, remove memory relationships panel
|
||||
merged_prs:
|
||||
- "#1247"
|
||||
|
||||
memory_birth:
|
||||
status: shipped
|
||||
files: [nexus/components/memory-birth.js]
|
||||
description: Birth animation when new memories are created
|
||||
merged_prs:
|
||||
- "#1222"
|
||||
|
||||
memory_particles:
|
||||
status: shipped
|
||||
files: [nexus/components/memory-particles.js]
|
||||
description: Ambient particle system — memory activity visualization
|
||||
merged_prs:
|
||||
- "#1205"
|
||||
|
||||
memory_optimizer:
|
||||
status: shipped
|
||||
files: [nexus/components/memory-optimizer.js]
|
||||
description: Performance optimization for large memory sets
|
||||
|
||||
timeline_scrubber:
|
||||
status: shipped
|
||||
files: [nexus/components/timeline-scrubber.js]
|
||||
description: Temporal navigation scrubber for memory timeline
|
||||
|
||||
health_dashboard:
|
||||
status: shipped
|
||||
files: [] # overlay in index.html
|
||||
description: Archive statistics overlay panel
|
||||
merged_prs:
|
||||
- "#1211"
|
||||
|
||||
# ─── Planned / Unshipped ──────────────────────────────────
|
||||
|
||||
planned:
|
||||
|
||||
memory_decay:
|
||||
status: shipped
|
||||
files: [entry.py, archive.py]
|
||||
description: >
|
||||
Memories have living energy that fades with neglect and
|
||||
brightens with access. Vitality score based on access
|
||||
frequency and recency. Exponential decay with 30-day half-life.
|
||||
Touch boost with diminishing returns.
|
||||
priority: medium
|
||||
merged_prs:
|
||||
- "#TBD" # Will be filled when PR is created
|
||||
|
||||
memory_pulse:
|
||||
status: shipped
|
||||
files: [nexus/components/memory-pulse.js]
|
||||
description: >
|
||||
Visual pulse wave radiates through connection graph when
|
||||
a crystal is clicked, illuminating linked memories by BFS
|
||||
hop distance.
|
||||
priority: medium
|
||||
merged_prs:
|
||||
- "#1263"
|
||||
|
||||
embedding_backend:
|
||||
status: shipped
|
||||
files: [embeddings.py]
|
||||
description: >
|
||||
Pluggable embedding backend for true semantic search.
|
||||
Supports Ollama (local models) and TF-IDF fallback.
|
||||
Auto-detects best available backend.
|
||||
priority: high
|
||||
merged_prs:
|
||||
- "#TBD" # Will be filled when PR is created
|
||||
|
||||
|
||||
memory_path:
|
||||
status: shipped
|
||||
files: [archive.py, cli.py, tests/test_path.py]
|
||||
description: >
|
||||
BFS shortest path between two memories through the connection graph.
|
||||
Answers "how is memory X related to memory Y?" by finding the chain
|
||||
of connections. Includes path_explanation for human-readable output.
|
||||
CLI command: mnemosyne path <start_id> <end_id>
|
||||
priority: medium
|
||||
merged_prs:
|
||||
- "#TBD"
|
||||
|
||||
memory_consolidation:
|
||||
status: shipped
|
||||
files: [archive.py, cli.py, tests/test_consolidation.py]
|
||||
description: >
|
||||
Automatic merging of duplicate/near-duplicate memories
|
||||
using content_hash and semantic similarity. Periodic
|
||||
consolidation pass.
|
||||
priority: low
|
||||
merged_prs:
|
||||
- "#1260"
|
||||
@@ -14,6 +14,12 @@ from nexus.mnemosyne.archive import MnemosyneArchive
|
||||
from nexus.mnemosyne.entry import ArchiveEntry
|
||||
from nexus.mnemosyne.linker import HolographicLinker
|
||||
from nexus.mnemosyne.ingest import ingest_from_mempalace, ingest_event
|
||||
from nexus.mnemosyne.embeddings import (
|
||||
EmbeddingBackend,
|
||||
OllamaEmbeddingBackend,
|
||||
TfidfEmbeddingBackend,
|
||||
get_embedding_backend,
|
||||
)
|
||||
|
||||
__all__ = [
|
||||
"MnemosyneArchive",
|
||||
@@ -21,4 +27,8 @@ __all__ = [
|
||||
"HolographicLinker",
|
||||
"ingest_from_mempalace",
|
||||
"ingest_event",
|
||||
"EmbeddingBackend",
|
||||
"OllamaEmbeddingBackend",
|
||||
"TfidfEmbeddingBackend",
|
||||
"get_embedding_backend",
|
||||
]
|
||||
|
||||
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
@@ -7,11 +7,13 @@ and provides query interfaces for retrieving connected knowledge.
|
||||
from __future__ import annotations
|
||||
|
||||
import json
|
||||
from datetime import datetime, timedelta, timezone
|
||||
from pathlib import Path
|
||||
from typing import Optional
|
||||
|
||||
from nexus.mnemosyne.entry import ArchiveEntry
|
||||
from nexus.mnemosyne.entry import ArchiveEntry, _compute_content_hash
|
||||
from nexus.mnemosyne.linker import HolographicLinker
|
||||
from nexus.mnemosyne.embeddings import get_embedding_backend, EmbeddingBackend
|
||||
|
||||
_EXPORT_VERSION = "1"
|
||||
|
||||
@@ -23,10 +25,21 @@ class MnemosyneArchive:
|
||||
MemPalace (ChromaDB) for vector-semantic search.
|
||||
"""
|
||||
|
||||
def __init__(self, archive_path: Optional[Path] = None):
|
||||
def __init__(
|
||||
self,
|
||||
archive_path: Optional[Path] = None,
|
||||
embedding_backend: Optional[EmbeddingBackend] = None,
|
||||
auto_embed: bool = True,
|
||||
):
|
||||
self.path = archive_path or Path.home() / ".hermes" / "mnemosyne" / "archive.json"
|
||||
self.path.parent.mkdir(parents=True, exist_ok=True)
|
||||
self.linker = HolographicLinker()
|
||||
self._embedding_backend = embedding_backend
|
||||
if embedding_backend is None and auto_embed:
|
||||
try:
|
||||
self._embedding_backend = get_embedding_backend()
|
||||
except Exception:
|
||||
self._embedding_backend = None
|
||||
self.linker = HolographicLinker(embedding_backend=self._embedding_backend)
|
||||
self._entries: dict[str, ArchiveEntry] = {}
|
||||
self._load()
|
||||
|
||||
@@ -49,28 +62,83 @@ class MnemosyneArchive:
|
||||
with open(self.path, "w") as f:
|
||||
json.dump(data, f, indent=2)
|
||||
|
||||
def add(self, entry: ArchiveEntry, auto_link: bool = True, skip_dups: bool = False) -> ArchiveEntry:
|
||||
def find_duplicate(self, entry: ArchiveEntry) -> Optional[ArchiveEntry]:
|
||||
"""Return an existing entry with the same content hash, or None."""
|
||||
for existing in self._entries.values():
|
||||
if existing.content_hash == entry.content_hash and existing.id != entry.id:
|
||||
return existing
|
||||
return None
|
||||
|
||||
def add(self, entry: ArchiveEntry, auto_link: bool = True) -> ArchiveEntry:
|
||||
"""Add an entry to the archive. Auto-links to related entries.
|
||||
|
||||
Args:
|
||||
entry: The entry to add.
|
||||
auto_link: Whether to automatically compute holographic links.
|
||||
skip_dups: If True, return existing entry instead of adding a duplicate
|
||||
(same title+content hash).
|
||||
|
||||
Returns:
|
||||
The added (or existing, if skip_dups=True and duplicate found) entry.
|
||||
If an entry with the same content hash already exists, returns the
|
||||
existing entry without creating a duplicate.
|
||||
"""
|
||||
if skip_dups:
|
||||
existing = self.find_by_hash(entry.content_hash)
|
||||
if existing:
|
||||
return existing
|
||||
duplicate = self.find_duplicate(entry)
|
||||
if duplicate is not None:
|
||||
return duplicate
|
||||
self._entries[entry.id] = entry
|
||||
if auto_link:
|
||||
self.linker.apply_links(entry, list(self._entries.values()))
|
||||
self._save()
|
||||
return entry
|
||||
|
||||
def update_entry(
|
||||
self,
|
||||
entry_id: str,
|
||||
title: Optional[str] = None,
|
||||
content: Optional[str] = None,
|
||||
metadata: Optional[dict] = None,
|
||||
auto_link: bool = True,
|
||||
) -> ArchiveEntry:
|
||||
"""Update title, content, and/or metadata on an existing entry.
|
||||
|
||||
Bumps ``updated_at`` and re-runs auto-linking when content changes.
|
||||
|
||||
Args:
|
||||
entry_id: ID of the entry to update.
|
||||
title: New title, or None to leave unchanged.
|
||||
content: New content, or None to leave unchanged.
|
||||
metadata: Dict to merge into existing metadata (replaces keys present).
|
||||
auto_link: If True, re-run holographic linker after content change.
|
||||
|
||||
Returns:
|
||||
The updated ArchiveEntry.
|
||||
|
||||
Raises:
|
||||
KeyError: If entry_id does not exist.
|
||||
"""
|
||||
entry = self._entries.get(entry_id)
|
||||
if entry is None:
|
||||
raise KeyError(entry_id)
|
||||
|
||||
content_changed = False
|
||||
if title is not None and title != entry.title:
|
||||
entry.title = title
|
||||
content_changed = True
|
||||
if content is not None and content != entry.content:
|
||||
entry.content = content
|
||||
content_changed = True
|
||||
if metadata is not None:
|
||||
entry.metadata.update(metadata)
|
||||
|
||||
if content_changed:
|
||||
entry.content_hash = _compute_content_hash(entry.title, entry.content)
|
||||
|
||||
entry.updated_at = datetime.now(timezone.utc).isoformat()
|
||||
|
||||
if content_changed and auto_link:
|
||||
# Clear old links from this entry and re-run linker
|
||||
for other in self._entries.values():
|
||||
if entry_id in other.links:
|
||||
other.links.remove(entry_id)
|
||||
entry.links = []
|
||||
self.linker.apply_links(entry, list(self._entries.values()))
|
||||
|
||||
self._save()
|
||||
return entry
|
||||
|
||||
def get(self, entry_id: str) -> Optional[ArchiveEntry]:
|
||||
return self._entries.get(entry_id)
|
||||
|
||||
@@ -87,33 +155,51 @@ class MnemosyneArchive:
|
||||
return [e for _, e in scored[:limit]]
|
||||
|
||||
def semantic_search(self, query: str, limit: int = 10, threshold: float = 0.05) -> list[ArchiveEntry]:
|
||||
"""Semantic search using holographic linker similarity.
|
||||
"""Semantic search using embeddings or holographic linker similarity.
|
||||
|
||||
Scores each entry by Jaccard similarity between query tokens and entry
|
||||
tokens, then boosts entries with more inbound links (more "holographic").
|
||||
Falls back to keyword search if no entries meet the similarity threshold.
|
||||
With an embedding backend: cosine similarity between query vector and
|
||||
entry vectors, boosted by inbound link count.
|
||||
Without: Jaccard similarity on tokens with link boost.
|
||||
Falls back to keyword search if nothing meets the threshold.
|
||||
|
||||
Args:
|
||||
query: Natural language query string.
|
||||
limit: Maximum number of results to return.
|
||||
threshold: Minimum Jaccard similarity to be considered a semantic match.
|
||||
threshold: Minimum similarity score to include in results.
|
||||
|
||||
Returns:
|
||||
List of ArchiveEntry sorted by combined relevance score, descending.
|
||||
"""
|
||||
query_tokens = HolographicLinker._tokenize(query)
|
||||
if not query_tokens:
|
||||
return []
|
||||
|
||||
# Count inbound links for each entry (how many entries link TO this one)
|
||||
# Count inbound links for link-boost
|
||||
inbound: dict[str, int] = {eid: 0 for eid in self._entries}
|
||||
for entry in self._entries.values():
|
||||
for linked_id in entry.links:
|
||||
if linked_id in inbound:
|
||||
inbound[linked_id] += 1
|
||||
|
||||
max_inbound = max(inbound.values(), default=1) or 1
|
||||
|
||||
# Try embedding-based search first
|
||||
if self._embedding_backend:
|
||||
query_vec = self._embedding_backend.embed(query)
|
||||
if query_vec:
|
||||
scored = []
|
||||
for entry in self._entries.values():
|
||||
text = f"{entry.title} {entry.content} {' '.join(entry.topics)}"
|
||||
entry_vec = self._embedding_backend.embed(text)
|
||||
if not entry_vec:
|
||||
continue
|
||||
sim = self._embedding_backend.similarity(query_vec, entry_vec)
|
||||
if sim >= threshold:
|
||||
link_boost = inbound[entry.id] / max_inbound * 0.15
|
||||
scored.append((sim + link_boost, entry))
|
||||
if scored:
|
||||
scored.sort(key=lambda x: x[0], reverse=True)
|
||||
return [e for _, e in scored[:limit]]
|
||||
|
||||
# Fallback: Jaccard token similarity
|
||||
query_tokens = HolographicLinker._tokenize(query)
|
||||
if not query_tokens:
|
||||
return []
|
||||
scored = []
|
||||
for entry in self._entries.values():
|
||||
entry_tokens = HolographicLinker._tokenize(f"{entry.title} {entry.content} {' '.join(entry.topics)}")
|
||||
@@ -123,14 +209,13 @@ class MnemosyneArchive:
|
||||
union = query_tokens | entry_tokens
|
||||
jaccard = len(intersection) / len(union)
|
||||
if jaccard >= threshold:
|
||||
link_boost = inbound[entry.id] / max_inbound * 0.2 # up to 20% boost
|
||||
link_boost = inbound[entry.id] / max_inbound * 0.2
|
||||
scored.append((jaccard + link_boost, entry))
|
||||
|
||||
if scored:
|
||||
scored.sort(key=lambda x: x[0], reverse=True)
|
||||
return [e for _, e in scored[:limit]]
|
||||
|
||||
# Graceful fallback to keyword search
|
||||
# Final fallback: keyword search
|
||||
return self.search(query, limit=limit)
|
||||
|
||||
def get_linked(self, entry_id: str, depth: int = 1) -> list[ArchiveEntry]:
|
||||
@@ -304,6 +389,17 @@ class MnemosyneArchive:
|
||||
oldest_entry = timestamps[0] if timestamps else None
|
||||
newest_entry = timestamps[-1] if timestamps else None
|
||||
|
||||
# Vitality summary
|
||||
if n > 0:
|
||||
vitalities = [self._compute_vitality(e) for e in entries]
|
||||
avg_vitality = round(sum(vitalities) / n, 4)
|
||||
fading_count = sum(1 for v in vitalities if v < 0.3)
|
||||
vibrant_count = sum(1 for v in vitalities if v > 0.7)
|
||||
else:
|
||||
avg_vitality = 0.0
|
||||
fading_count = 0
|
||||
vibrant_count = 0
|
||||
|
||||
return {
|
||||
"entries": n,
|
||||
"total_links": total_links,
|
||||
@@ -313,6 +409,9 @@ class MnemosyneArchive:
|
||||
"link_density": link_density,
|
||||
"oldest_entry": oldest_entry,
|
||||
"newest_entry": newest_entry,
|
||||
"avg_vitality": avg_vitality,
|
||||
"fading_count": fading_count,
|
||||
"vibrant_count": vibrant_count,
|
||||
}
|
||||
|
||||
def _build_adjacency(self) -> dict[str, set[str]]:
|
||||
@@ -595,25 +694,83 @@ class MnemosyneArchive:
|
||||
self._save()
|
||||
return entry
|
||||
|
||||
def update_entry(
|
||||
self,
|
||||
entry_id: str,
|
||||
title: Optional[str] = None,
|
||||
content: Optional[str] = None,
|
||||
metadata: Optional[dict] = None,
|
||||
re_link: bool = True,
|
||||
) -> ArchiveEntry:
|
||||
"""Update fields on an existing entry.
|
||||
@staticmethod
|
||||
def _parse_dt(dt_str: str) -> datetime:
|
||||
"""Parse an ISO datetime string. Assumes UTC if no timezone is specified."""
|
||||
dt = datetime.fromisoformat(dt_str)
|
||||
if dt.tzinfo is None:
|
||||
dt = dt.replace(tzinfo=timezone.utc)
|
||||
return dt
|
||||
|
||||
Only provided fields are changed. Bumps updated_at and optionally
|
||||
recomputes holographic links (since content changed).
|
||||
def by_date_range(self, start: str, end: str) -> list[ArchiveEntry]:
|
||||
"""Return entries whose ``created_at`` falls within [start, end] (inclusive).
|
||||
|
||||
Args:
|
||||
entry_id: ID of the entry to update.
|
||||
title: New title (None = keep existing).
|
||||
content: New content (None = keep existing).
|
||||
metadata: New metadata dict (None = keep existing, {} to clear).
|
||||
re_link: Whether to recompute holographic links after update.
|
||||
start: ISO datetime string for the range start (e.g. "2024-01-01" or
|
||||
"2024-01-01T00:00:00Z"). Timezone-naive strings are treated as UTC.
|
||||
end: ISO datetime string for the range end. Timezone-naive strings are
|
||||
treated as UTC.
|
||||
|
||||
Returns:
|
||||
List of ArchiveEntry sorted by ``created_at`` ascending.
|
||||
"""
|
||||
start_dt = self._parse_dt(start)
|
||||
end_dt = self._parse_dt(end)
|
||||
results = []
|
||||
for entry in self._entries.values():
|
||||
entry_dt = self._parse_dt(entry.created_at)
|
||||
if start_dt <= entry_dt <= end_dt:
|
||||
results.append(entry)
|
||||
results.sort(key=lambda e: e.created_at)
|
||||
return results
|
||||
|
||||
def temporal_neighbors(self, entry_id: str, window_days: int = 7) -> list[ArchiveEntry]:
|
||||
"""Return entries created within ``window_days`` of a given entry.
|
||||
|
||||
The reference entry itself is excluded from results.
|
||||
|
||||
Args:
|
||||
entry_id: ID of the anchor entry.
|
||||
window_days: Number of days around the anchor's ``created_at`` to search.
|
||||
|
||||
Returns:
|
||||
List of ArchiveEntry sorted by ``created_at`` ascending.
|
||||
|
||||
Raises:
|
||||
KeyError: If ``entry_id`` does not exist in the archive.
|
||||
"""
|
||||
anchor = self._entries.get(entry_id)
|
||||
if anchor is None:
|
||||
raise KeyError(entry_id)
|
||||
anchor_dt = self._parse_dt(anchor.created_at)
|
||||
delta = timedelta(days=window_days)
|
||||
window_start = anchor_dt - delta
|
||||
window_end = anchor_dt + delta
|
||||
results = []
|
||||
for entry in self._entries.values():
|
||||
if entry.id == entry_id:
|
||||
continue
|
||||
entry_dt = self._parse_dt(entry.created_at)
|
||||
if window_start <= entry_dt <= window_end:
|
||||
results.append(entry)
|
||||
results.sort(key=lambda e: e.created_at)
|
||||
return results
|
||||
|
||||
# ─── Memory Decay ─────────────────────────────────────────
|
||||
|
||||
# Decay parameters
|
||||
_DECAY_HALF_LIFE_DAYS: float = 30.0 # Half-life for exponential decay
|
||||
_TOUCH_BOOST_FACTOR: float = 0.1 # Base boost on access (diminishes as vitality → 1.0)
|
||||
|
||||
def touch(self, entry_id: str) -> ArchiveEntry:
|
||||
"""Record an access to an entry, boosting its vitality.
|
||||
|
||||
The boost is ``_TOUCH_BOOST_FACTOR * (1 - current_vitality)`` —
|
||||
diminishing returns as vitality approaches 1.0 ensures entries
|
||||
can never exceed 1.0 through touch alone.
|
||||
|
||||
Args:
|
||||
entry_id: ID of the entry to touch.
|
||||
|
||||
Returns:
|
||||
The updated ArchiveEntry.
|
||||
@@ -625,52 +782,563 @@ class MnemosyneArchive:
|
||||
if entry is None:
|
||||
raise KeyError(entry_id)
|
||||
|
||||
old_hash = entry.content_hash
|
||||
|
||||
if title is not None:
|
||||
entry.title = title
|
||||
if content is not None:
|
||||
entry.content = content
|
||||
if metadata is not None:
|
||||
entry.metadata = metadata
|
||||
entry.touch()
|
||||
|
||||
# Re-link only if content actually changed
|
||||
if re_link and entry.content_hash != old_hash:
|
||||
# Clear existing links to this entry from others
|
||||
for other in self._entries.values():
|
||||
if entry_id in other.links:
|
||||
other.links.remove(entry_id)
|
||||
entry.links = []
|
||||
# Re-apply
|
||||
self.linker.apply_links(entry, list(self._entries.values()))
|
||||
now = datetime.now(timezone.utc).isoformat()
|
||||
|
||||
# Compute current decayed vitality before boosting
|
||||
current = self._compute_vitality(entry)
|
||||
boost = self._TOUCH_BOOST_FACTOR * (1.0 - current)
|
||||
entry.vitality = min(1.0, current + boost)
|
||||
entry.last_accessed = now
|
||||
self._save()
|
||||
return entry
|
||||
|
||||
def find_by_hash(self, content_hash: str) -> Optional[ArchiveEntry]:
|
||||
"""Find an entry by its content hash (title + content SHA-256).
|
||||
def _compute_vitality(self, entry: ArchiveEntry) -> float:
|
||||
"""Compute the current vitality of an entry based on time decay.
|
||||
|
||||
Returns the first match, or None if no entry has this hash.
|
||||
Uses exponential decay: ``v = base * 0.5 ^ (hours_since_access / half_life_hours)``
|
||||
|
||||
If the entry has never been accessed, uses ``created_at`` as the
|
||||
reference point. New entries with no access start at full vitality.
|
||||
|
||||
Args:
|
||||
entry: The archive entry.
|
||||
|
||||
Returns:
|
||||
Current vitality as a float in [0.0, 1.0].
|
||||
"""
|
||||
if entry.last_accessed is None:
|
||||
# Never accessed — check age from creation
|
||||
created = self._parse_dt(entry.created_at)
|
||||
hours_elapsed = (datetime.now(timezone.utc) - created).total_seconds() / 3600
|
||||
else:
|
||||
last = self._parse_dt(entry.last_accessed)
|
||||
hours_elapsed = (datetime.now(timezone.utc) - last).total_seconds() / 3600
|
||||
|
||||
half_life_hours = self._DECAY_HALF_LIFE_DAYS * 24
|
||||
if hours_elapsed <= 0 or half_life_hours <= 0:
|
||||
return entry.vitality
|
||||
|
||||
decayed = entry.vitality * (0.5 ** (hours_elapsed / half_life_hours))
|
||||
return max(0.0, min(1.0, decayed))
|
||||
|
||||
def get_vitality(self, entry_id: str) -> dict:
|
||||
"""Get the current vitality status of an entry.
|
||||
|
||||
Args:
|
||||
entry_id: ID of the entry.
|
||||
|
||||
Returns:
|
||||
Dict with keys: entry_id, title, vitality, last_accessed, age_days
|
||||
|
||||
Raises:
|
||||
KeyError: If entry_id does not exist.
|
||||
"""
|
||||
entry = self._entries.get(entry_id)
|
||||
if entry is None:
|
||||
raise KeyError(entry_id)
|
||||
|
||||
current_vitality = self._compute_vitality(entry)
|
||||
created = self._parse_dt(entry.created_at)
|
||||
age_days = (datetime.now(timezone.utc) - created).days
|
||||
|
||||
return {
|
||||
"entry_id": entry.id,
|
||||
"title": entry.title,
|
||||
"vitality": round(current_vitality, 4),
|
||||
"last_accessed": entry.last_accessed,
|
||||
"age_days": age_days,
|
||||
}
|
||||
|
||||
def fading(self, limit: int = 10) -> list[dict]:
|
||||
"""Return entries with the lowest vitality (most neglected).
|
||||
|
||||
Args:
|
||||
limit: Maximum number of entries to return.
|
||||
|
||||
Returns:
|
||||
List of dicts sorted by vitality ascending (most faded first).
|
||||
Each dict has keys: entry_id, title, vitality, last_accessed, age_days
|
||||
"""
|
||||
scored = []
|
||||
for entry in self._entries.values():
|
||||
if entry.content_hash == content_hash:
|
||||
return entry
|
||||
v = self._compute_vitality(entry)
|
||||
created = self._parse_dt(entry.created_at)
|
||||
age_days = (datetime.now(timezone.utc) - created).days
|
||||
scored.append({
|
||||
"entry_id": entry.id,
|
||||
"title": entry.title,
|
||||
"vitality": round(v, 4),
|
||||
"last_accessed": entry.last_accessed,
|
||||
"age_days": age_days,
|
||||
})
|
||||
scored.sort(key=lambda x: x["vitality"])
|
||||
return scored[:limit]
|
||||
|
||||
def vibrant(self, limit: int = 10) -> list[dict]:
|
||||
"""Return entries with the highest vitality (most alive).
|
||||
|
||||
Args:
|
||||
limit: Maximum number of entries to return.
|
||||
|
||||
Returns:
|
||||
List of dicts sorted by vitality descending (most vibrant first).
|
||||
Each dict has keys: entry_id, title, vitality, last_accessed, age_days
|
||||
"""
|
||||
scored = []
|
||||
for entry in self._entries.values():
|
||||
v = self._compute_vitality(entry)
|
||||
created = self._parse_dt(entry.created_at)
|
||||
age_days = (datetime.now(timezone.utc) - created).days
|
||||
scored.append({
|
||||
"entry_id": entry.id,
|
||||
"title": entry.title,
|
||||
"vitality": round(v, 4),
|
||||
"last_accessed": entry.last_accessed,
|
||||
"age_days": age_days,
|
||||
})
|
||||
scored.sort(key=lambda x: x["vitality"], reverse=True)
|
||||
return scored[:limit]
|
||||
|
||||
def apply_decay(self) -> dict:
|
||||
"""Apply time-based decay to all entries and persist.
|
||||
|
||||
Recomputes each entry's vitality based on elapsed time since
|
||||
its last access (or creation if never accessed). Saves the
|
||||
archive after updating.
|
||||
|
||||
Returns:
|
||||
Dict with keys: total_entries, decayed_count, avg_vitality,
|
||||
fading_count (entries below 0.3), vibrant_count (entries above 0.7)
|
||||
"""
|
||||
decayed = 0
|
||||
total_vitality = 0.0
|
||||
fading_count = 0
|
||||
vibrant_count = 0
|
||||
|
||||
for entry in self._entries.values():
|
||||
old_v = entry.vitality
|
||||
new_v = self._compute_vitality(entry)
|
||||
if abs(new_v - old_v) > 1e-6:
|
||||
entry.vitality = new_v
|
||||
decayed += 1
|
||||
total_vitality += entry.vitality
|
||||
if entry.vitality < 0.3:
|
||||
fading_count += 1
|
||||
if entry.vitality > 0.7:
|
||||
vibrant_count += 1
|
||||
|
||||
n = len(self._entries)
|
||||
self._save()
|
||||
|
||||
return {
|
||||
"total_entries": n,
|
||||
"decayed_count": decayed,
|
||||
"avg_vitality": round(total_vitality / n, 4) if n else 0.0,
|
||||
"fading_count": fading_count,
|
||||
"vibrant_count": vibrant_count,
|
||||
}
|
||||
|
||||
def consolidate(
|
||||
self,
|
||||
threshold: float = 0.9,
|
||||
dry_run: bool = False,
|
||||
) -> list[dict]:
|
||||
"""Scan the archive and merge duplicate/near-duplicate entries.
|
||||
|
||||
Two entries are considered duplicates if:
|
||||
- They share the same ``content_hash`` (exact duplicate), or
|
||||
- Their similarity score (via HolographicLinker) exceeds ``threshold``
|
||||
(near-duplicate when an embedding backend is available or Jaccard is
|
||||
high enough at the given threshold).
|
||||
|
||||
Merge strategy:
|
||||
- Keep the *older* entry (earlier ``created_at``).
|
||||
- Union topics from both entries (case-deduped).
|
||||
- Merge metadata from newer into older (older values win on conflicts).
|
||||
- Transfer all links from the newer entry to the older entry.
|
||||
- Delete the newer entry.
|
||||
|
||||
Args:
|
||||
threshold: Similarity threshold for near-duplicate detection (0.0–1.0).
|
||||
Default 0.9 is intentionally conservative.
|
||||
dry_run: If True, return the list of would-be merges without mutating
|
||||
the archive.
|
||||
|
||||
Returns:
|
||||
List of dicts, one per merged pair::
|
||||
|
||||
{
|
||||
"kept": <entry_id of survivor>,
|
||||
"removed": <entry_id of duplicate>,
|
||||
"reason": "exact_hash" | "semantic_similarity",
|
||||
"score": float, # 1.0 for exact hash matches
|
||||
"dry_run": bool,
|
||||
}
|
||||
"""
|
||||
merges: list[dict] = []
|
||||
entries = list(self._entries.values())
|
||||
removed_ids: set[str] = set()
|
||||
|
||||
for i, entry_a in enumerate(entries):
|
||||
if entry_a.id in removed_ids:
|
||||
continue
|
||||
for entry_b in entries[i + 1:]:
|
||||
if entry_b.id in removed_ids:
|
||||
continue
|
||||
|
||||
# Determine if they are duplicates
|
||||
reason: Optional[str] = None
|
||||
score: float = 0.0
|
||||
|
||||
if (
|
||||
entry_a.content_hash is not None
|
||||
and entry_b.content_hash is not None
|
||||
and entry_a.content_hash == entry_b.content_hash
|
||||
):
|
||||
reason = "exact_hash"
|
||||
score = 1.0
|
||||
else:
|
||||
sim = self.linker.compute_similarity(entry_a, entry_b)
|
||||
if sim >= threshold:
|
||||
reason = "semantic_similarity"
|
||||
score = sim
|
||||
|
||||
if reason is None:
|
||||
continue
|
||||
|
||||
# Decide which entry to keep (older survives)
|
||||
if entry_a.created_at <= entry_b.created_at:
|
||||
kept, removed = entry_a, entry_b
|
||||
else:
|
||||
kept, removed = entry_b, entry_a
|
||||
|
||||
merges.append({
|
||||
"kept": kept.id,
|
||||
"removed": removed.id,
|
||||
"reason": reason,
|
||||
"score": round(score, 4),
|
||||
"dry_run": dry_run,
|
||||
})
|
||||
|
||||
if not dry_run:
|
||||
# Merge topics (case-deduped)
|
||||
existing_lower = {t.lower() for t in kept.topics}
|
||||
for tag in removed.topics:
|
||||
if tag.lower() not in existing_lower:
|
||||
kept.topics.append(tag)
|
||||
existing_lower.add(tag.lower())
|
||||
|
||||
# Merge metadata (kept wins on key conflicts)
|
||||
for k, v in removed.metadata.items():
|
||||
if k not in kept.metadata:
|
||||
kept.metadata[k] = v
|
||||
|
||||
# Transfer links: add removed's links to kept
|
||||
kept_links_set = set(kept.links)
|
||||
for lid in removed.links:
|
||||
if lid != kept.id and lid not in kept_links_set and lid not in removed_ids:
|
||||
kept.links.append(lid)
|
||||
kept_links_set.add(lid)
|
||||
# Update the other entry's back-link
|
||||
other = self._entries.get(lid)
|
||||
if other and kept.id not in other.links:
|
||||
other.links.append(kept.id)
|
||||
|
||||
# Remove back-links pointing at the removed entry
|
||||
for other in self._entries.values():
|
||||
if removed.id in other.links:
|
||||
other.links.remove(removed.id)
|
||||
if other.id != kept.id and kept.id not in other.links:
|
||||
other.links.append(kept.id)
|
||||
|
||||
del self._entries[removed.id]
|
||||
removed_ids.add(removed.id)
|
||||
|
||||
if not dry_run and merges:
|
||||
self._save()
|
||||
|
||||
return merges
|
||||
|
||||
|
||||
def shortest_path(self, start_id: str, end_id: str) -> list[str] | None:
|
||||
"""Find shortest path between two entries through the connection graph.
|
||||
|
||||
Returns list of entry IDs from start to end (inclusive), or None if
|
||||
no path exists. Uses BFS for unweighted shortest path.
|
||||
"""
|
||||
if start_id == end_id:
|
||||
return [start_id] if start_id in self._entries else None
|
||||
if start_id not in self._entries or end_id not in self._entries:
|
||||
return None
|
||||
|
||||
adj = self._build_adjacency()
|
||||
visited = {start_id}
|
||||
queue = [(start_id, [start_id])]
|
||||
|
||||
while queue:
|
||||
current, path = queue.pop(0)
|
||||
for neighbor in adj.get(current, []):
|
||||
if neighbor == end_id:
|
||||
return path + [neighbor]
|
||||
if neighbor not in visited:
|
||||
visited.add(neighbor)
|
||||
queue.append((neighbor, path + [neighbor]))
|
||||
|
||||
return None
|
||||
|
||||
def find_duplicates(self) -> list[list[ArchiveEntry]]:
|
||||
"""Find groups of entries with identical content hashes.
|
||||
def path_explanation(self, path: list[str]) -> list[dict]:
|
||||
"""Convert a path of entry IDs into human-readable step descriptions.
|
||||
|
||||
Returns a list of groups, where each group is a list of 2+ entries
|
||||
sharing the same title+content. Sorted by group size descending.
|
||||
Returns list of dicts with 'id', 'title', and 'topics' for each step.
|
||||
"""
|
||||
hash_groups: dict[str, list[ArchiveEntry]] = {}
|
||||
for entry in self._entries.values():
|
||||
h = entry.content_hash
|
||||
hash_groups.setdefault(h, []).append(entry)
|
||||
dups = [group for group in hash_groups.values() if len(group) > 1]
|
||||
dups.sort(key=lambda g: len(g), reverse=True)
|
||||
return dups
|
||||
steps = []
|
||||
for entry_id in path:
|
||||
entry = self._entries.get(entry_id)
|
||||
if entry:
|
||||
steps.append({
|
||||
"id": entry.id,
|
||||
"title": entry.title,
|
||||
"topics": entry.topics,
|
||||
"content_preview": entry.content[:120] + "..." if len(entry.content) > 120 else entry.content,
|
||||
})
|
||||
else:
|
||||
steps.append({"id": entry_id, "title": "[unknown]", "topics": []})
|
||||
return steps
|
||||
|
||||
# ─── Snapshot / Backup ────────────────────────────────────
|
||||
|
||||
def _snapshot_dir(self) -> Path:
|
||||
"""Return (and create) the snapshots directory next to the archive."""
|
||||
d = self.path.parent / "snapshots"
|
||||
d.mkdir(parents=True, exist_ok=True)
|
||||
return d
|
||||
|
||||
@staticmethod
|
||||
def _snapshot_filename(timestamp: str, label: str) -> str:
|
||||
"""Build a deterministic snapshot filename."""
|
||||
safe_label = "".join(c if c.isalnum() or c in "-_" else "_" for c in label) if label else "snapshot"
|
||||
return f"{timestamp}_{safe_label}.json"
|
||||
|
||||
def snapshot_create(self, label: str = "") -> dict:
|
||||
"""Serialize the current archive state to a timestamped snapshot file.
|
||||
|
||||
Args:
|
||||
label: Human-readable label for the snapshot (optional).
|
||||
|
||||
Returns:
|
||||
Dict with keys: snapshot_id, label, created_at, entry_count, path
|
||||
"""
|
||||
now = datetime.now(timezone.utc)
|
||||
timestamp = now.strftime("%Y%m%d_%H%M%S")
|
||||
filename = self._snapshot_filename(timestamp, label)
|
||||
snapshot_id = filename[:-5] # strip .json
|
||||
snap_path = self._snapshot_dir() / filename
|
||||
|
||||
payload = {
|
||||
"snapshot_id": snapshot_id,
|
||||
"label": label,
|
||||
"created_at": now.isoformat(),
|
||||
"entry_count": len(self._entries),
|
||||
"archive_path": str(self.path),
|
||||
"entries": [e.to_dict() for e in self._entries.values()],
|
||||
}
|
||||
with open(snap_path, "w") as f:
|
||||
json.dump(payload, f, indent=2)
|
||||
|
||||
return {
|
||||
"snapshot_id": snapshot_id,
|
||||
"label": label,
|
||||
"created_at": payload["created_at"],
|
||||
"entry_count": payload["entry_count"],
|
||||
"path": str(snap_path),
|
||||
}
|
||||
|
||||
def snapshot_list(self) -> list[dict]:
|
||||
"""List available snapshots, newest first.
|
||||
|
||||
Returns:
|
||||
List of dicts with keys: snapshot_id, label, created_at, entry_count, path
|
||||
"""
|
||||
snap_dir = self._snapshot_dir()
|
||||
snapshots = []
|
||||
for snap_path in sorted(snap_dir.glob("*.json"), reverse=True):
|
||||
try:
|
||||
with open(snap_path) as f:
|
||||
data = json.load(f)
|
||||
snapshots.append({
|
||||
"snapshot_id": data.get("snapshot_id", snap_path.stem),
|
||||
"label": data.get("label", ""),
|
||||
"created_at": data.get("created_at", ""),
|
||||
"entry_count": data.get("entry_count", len(data.get("entries", []))),
|
||||
"path": str(snap_path),
|
||||
})
|
||||
except (json.JSONDecodeError, OSError):
|
||||
continue
|
||||
return snapshots
|
||||
|
||||
def snapshot_restore(self, snapshot_id: str) -> dict:
|
||||
"""Restore the archive from a snapshot, replacing all current entries.
|
||||
|
||||
Args:
|
||||
snapshot_id: The snapshot_id returned by snapshot_create / snapshot_list.
|
||||
|
||||
Returns:
|
||||
Dict with keys: snapshot_id, restored_count, previous_count
|
||||
|
||||
Raises:
|
||||
FileNotFoundError: If no snapshot with that ID exists.
|
||||
"""
|
||||
snap_dir = self._snapshot_dir()
|
||||
snap_path = snap_dir / f"{snapshot_id}.json"
|
||||
if not snap_path.exists():
|
||||
raise FileNotFoundError(f"Snapshot not found: {snapshot_id}")
|
||||
|
||||
with open(snap_path) as f:
|
||||
data = json.load(f)
|
||||
|
||||
previous_count = len(self._entries)
|
||||
self._entries = {}
|
||||
for entry_data in data.get("entries", []):
|
||||
entry = ArchiveEntry.from_dict(entry_data)
|
||||
self._entries[entry.id] = entry
|
||||
|
||||
self._save()
|
||||
return {
|
||||
"snapshot_id": snapshot_id,
|
||||
"restored_count": len(self._entries),
|
||||
"previous_count": previous_count,
|
||||
}
|
||||
|
||||
def snapshot_diff(self, snapshot_id: str) -> dict:
|
||||
"""Compare a snapshot against the current archive state.
|
||||
|
||||
Args:
|
||||
snapshot_id: The snapshot_id to compare against current state.
|
||||
|
||||
Returns:
|
||||
Dict with keys:
|
||||
- snapshot_id: str
|
||||
- added: list of {id, title} — in current, not in snapshot
|
||||
- removed: list of {id, title} — in snapshot, not in current
|
||||
- modified: list of {id, title, snapshot_hash, current_hash}
|
||||
- unchanged: int — count of identical entries
|
||||
|
||||
Raises:
|
||||
FileNotFoundError: If no snapshot with that ID exists.
|
||||
"""
|
||||
snap_dir = self._snapshot_dir()
|
||||
snap_path = snap_dir / f"{snapshot_id}.json"
|
||||
if not snap_path.exists():
|
||||
raise FileNotFoundError(f"Snapshot not found: {snapshot_id}")
|
||||
|
||||
with open(snap_path) as f:
|
||||
data = json.load(f)
|
||||
|
||||
snap_entries: dict[str, dict] = {}
|
||||
for entry_data in data.get("entries", []):
|
||||
snap_entries[entry_data["id"]] = entry_data
|
||||
|
||||
current_ids = set(self._entries.keys())
|
||||
snap_ids = set(snap_entries.keys())
|
||||
|
||||
added = []
|
||||
for eid in current_ids - snap_ids:
|
||||
e = self._entries[eid]
|
||||
added.append({"id": e.id, "title": e.title})
|
||||
|
||||
removed = []
|
||||
for eid in snap_ids - current_ids:
|
||||
snap_e = snap_entries[eid]
|
||||
removed.append({"id": snap_e["id"], "title": snap_e.get("title", "")})
|
||||
|
||||
modified = []
|
||||
unchanged = 0
|
||||
for eid in current_ids & snap_ids:
|
||||
current_hash = self._entries[eid].content_hash
|
||||
snap_hash = snap_entries[eid].get("content_hash")
|
||||
if current_hash != snap_hash:
|
||||
modified.append({
|
||||
"id": eid,
|
||||
"title": self._entries[eid].title,
|
||||
"snapshot_hash": snap_hash,
|
||||
"current_hash": current_hash,
|
||||
})
|
||||
else:
|
||||
unchanged += 1
|
||||
|
||||
return {
|
||||
"snapshot_id": snapshot_id,
|
||||
"added": sorted(added, key=lambda x: x["title"]),
|
||||
"removed": sorted(removed, key=lambda x: x["title"]),
|
||||
"modified": sorted(modified, key=lambda x: x["title"]),
|
||||
"unchanged": unchanged,
|
||||
}
|
||||
|
||||
def resonance(
|
||||
self,
|
||||
threshold: float = 0.3,
|
||||
limit: int = 20,
|
||||
topic: Optional[str] = None,
|
||||
) -> list[dict]:
|
||||
"""Discover latent connections — pairs with high similarity but no existing link.
|
||||
|
||||
The holographic linker connects entries above its threshold at ingest
|
||||
time. ``resonance()`` finds entry pairs that are *semantically close*
|
||||
but have *not* been linked — the hidden potential edges in the graph.
|
||||
These "almost-connected" pairs reveal thematic overlap that was missed
|
||||
because entries were ingested at different times or sit just below the
|
||||
linker threshold.
|
||||
|
||||
Args:
|
||||
threshold: Minimum similarity score to surface a pair (default 0.3).
|
||||
Pairs already linked are excluded regardless of score.
|
||||
limit: Maximum number of pairs to return (default 20).
|
||||
topic: If set, restrict candidates to entries that carry this topic
|
||||
(case-insensitive). Both entries in a pair must match.
|
||||
|
||||
Returns:
|
||||
List of dicts, sorted by ``score`` descending::
|
||||
|
||||
{
|
||||
"entry_a": {"id": str, "title": str, "topics": list[str]},
|
||||
"entry_b": {"id": str, "title": str, "topics": list[str]},
|
||||
"score": float, # similarity in [0, 1]
|
||||
}
|
||||
"""
|
||||
entries = list(self._entries.values())
|
||||
|
||||
if topic:
|
||||
topic_lower = topic.lower()
|
||||
entries = [e for e in entries if topic_lower in [t.lower() for t in e.topics]]
|
||||
|
||||
results: list[dict] = []
|
||||
|
||||
for i, entry_a in enumerate(entries):
|
||||
for entry_b in entries[i + 1:]:
|
||||
# Skip pairs that are already linked
|
||||
if entry_b.id in entry_a.links or entry_a.id in entry_b.links:
|
||||
continue
|
||||
|
||||
score = self.linker.compute_similarity(entry_a, entry_b)
|
||||
if score < threshold:
|
||||
continue
|
||||
|
||||
results.append({
|
||||
"entry_a": {
|
||||
"id": entry_a.id,
|
||||
"title": entry_a.title,
|
||||
"topics": entry_a.topics,
|
||||
},
|
||||
"entry_b": {
|
||||
"id": entry_b.id,
|
||||
"title": entry_b.title,
|
||||
"topics": entry_b.topics,
|
||||
},
|
||||
"score": round(score, 4),
|
||||
})
|
||||
|
||||
results.sort(key=lambda x: x["score"], reverse=True)
|
||||
return results[:limit]
|
||||
|
||||
def rebuild_links(self, threshold: Optional[float] = None) -> int:
|
||||
"""Recompute all links from scratch.
|
||||
|
||||
@@ -3,7 +3,12 @@
|
||||
Provides: mnemosyne ingest, mnemosyne search, mnemosyne link, mnemosyne stats,
|
||||
mnemosyne topics, mnemosyne remove, mnemosyne export,
|
||||
mnemosyne clusters, mnemosyne hubs, mnemosyne bridges, mnemosyne rebuild,
|
||||
mnemosyne tag, mnemosyne untag, mnemosyne retag
|
||||
mnemosyne tag, mnemosyne untag, mnemosyne retag,
|
||||
mnemosyne timeline, mnemosyne neighbors, mnemosyne path,
|
||||
mnemosyne touch, mnemosyne decay, mnemosyne vitality,
|
||||
mnemosyne fading, mnemosyne vibrant,
|
||||
mnemosyne snapshot create|list|restore|diff,
|
||||
mnemosyne resonance
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
@@ -14,7 +19,7 @@ import sys
|
||||
|
||||
from nexus.mnemosyne.archive import MnemosyneArchive
|
||||
from nexus.mnemosyne.entry import ArchiveEntry
|
||||
from nexus.mnemosyne.ingest import ingest_event
|
||||
from nexus.mnemosyne.ingest import ingest_event, ingest_directory
|
||||
|
||||
|
||||
def cmd_stats(args):
|
||||
@@ -24,7 +29,16 @@ def cmd_stats(args):
|
||||
|
||||
|
||||
def cmd_search(args):
|
||||
archive = MnemosyneArchive()
|
||||
from nexus.mnemosyne.embeddings import get_embedding_backend
|
||||
backend = None
|
||||
if getattr(args, "backend", "auto") != "auto":
|
||||
backend = get_embedding_backend(prefer=args.backend)
|
||||
elif getattr(args, "semantic", False):
|
||||
try:
|
||||
backend = get_embedding_backend()
|
||||
except Exception:
|
||||
pass
|
||||
archive = MnemosyneArchive(embedding_backend=backend)
|
||||
if getattr(args, "semantic", False):
|
||||
results = archive.semantic_search(args.query, limit=args.limit)
|
||||
else:
|
||||
@@ -51,6 +65,13 @@ def cmd_ingest(args):
|
||||
print(f"Ingested: [{entry.id[:8]}] {entry.title} ({len(entry.links)} links)")
|
||||
|
||||
|
||||
def cmd_ingest_dir(args):
|
||||
archive = MnemosyneArchive()
|
||||
ext = [e.strip() for e in args.ext.split(",")] if args.ext else None
|
||||
added = ingest_directory(archive, args.path, extensions=ext)
|
||||
print(f"Ingested {added} new entries from {args.path}")
|
||||
|
||||
|
||||
def cmd_link(args):
|
||||
archive = MnemosyneArchive()
|
||||
entry = archive.get(args.entry_id)
|
||||
@@ -180,6 +201,209 @@ def cmd_retag(args):
|
||||
print(f" Topics: {', '.join(entry.topics) if entry.topics else '(none)'}")
|
||||
|
||||
|
||||
def cmd_timeline(args):
|
||||
archive = MnemosyneArchive()
|
||||
try:
|
||||
results = archive.by_date_range(args.start, args.end)
|
||||
except ValueError as e:
|
||||
print(f"Invalid date format: {e}")
|
||||
sys.exit(1)
|
||||
if not results:
|
||||
print("No entries found in that date range.")
|
||||
return
|
||||
for entry in results:
|
||||
print(f"[{entry.id[:8]}] {entry.created_at[:10]} {entry.title}")
|
||||
print(f" Topics: {', '.join(entry.topics) if entry.topics else '(none)'}")
|
||||
print()
|
||||
|
||||
|
||||
|
||||
def cmd_path(args):
|
||||
archive = MnemosyneArchive(archive_path=args.archive) if args.archive else MnemosyneArchive()
|
||||
path = archive.shortest_path(args.start, args.end)
|
||||
if path is None:
|
||||
print(f"No path found between {args.start} and {args.end}")
|
||||
return
|
||||
steps = archive.path_explanation(path)
|
||||
print(f"Path ({len(steps)} hops):")
|
||||
for i, step in enumerate(steps):
|
||||
arrow = " → " if i > 0 else " "
|
||||
print(f"{arrow}{step['id']}: {step['title']}")
|
||||
if step['topics']:
|
||||
print(f" topics: {', '.join(step['topics'])}")
|
||||
|
||||
def cmd_consolidate(args):
|
||||
archive = MnemosyneArchive()
|
||||
merges = archive.consolidate(threshold=args.threshold, dry_run=args.dry_run)
|
||||
if not merges:
|
||||
print("No duplicates found.")
|
||||
return
|
||||
label = "[DRY RUN] " if args.dry_run else ""
|
||||
for m in merges:
|
||||
print(f"{label}Merge ({m['reason']}, score={m['score']:.4f}):")
|
||||
print(f" kept: {m['kept'][:8]}")
|
||||
print(f" removed: {m['removed'][:8]}")
|
||||
if args.dry_run:
|
||||
print(f"\n{len(merges)} pair(s) would be merged. Re-run without --dry-run to apply.")
|
||||
else:
|
||||
print(f"\nMerged {len(merges)} duplicate pair(s).")
|
||||
|
||||
|
||||
def cmd_neighbors(args):
|
||||
archive = MnemosyneArchive()
|
||||
try:
|
||||
results = archive.temporal_neighbors(args.entry_id, window_days=args.days)
|
||||
except KeyError:
|
||||
print(f"Entry not found: {args.entry_id}")
|
||||
sys.exit(1)
|
||||
if not results:
|
||||
print("No temporal neighbors found.")
|
||||
return
|
||||
for entry in results:
|
||||
print(f"[{entry.id[:8]}] {entry.created_at[:10]} {entry.title}")
|
||||
print(f" Topics: {', '.join(entry.topics) if entry.topics else '(none)'}")
|
||||
print()
|
||||
|
||||
|
||||
def cmd_touch(args):
|
||||
archive = MnemosyneArchive()
|
||||
try:
|
||||
entry = archive.touch(args.entry_id)
|
||||
except KeyError:
|
||||
print(f"Entry not found: {args.entry_id}")
|
||||
sys.exit(1)
|
||||
v = archive.get_vitality(entry.id)
|
||||
print(f"[{entry.id[:8]}] {entry.title}")
|
||||
print(f" Vitality: {v['vitality']:.4f} (boosted)")
|
||||
|
||||
|
||||
def cmd_decay(args):
|
||||
archive = MnemosyneArchive()
|
||||
result = archive.apply_decay()
|
||||
print(f"Applied decay to {result['total_entries']} entries")
|
||||
print(f" Decayed: {result['decayed_count']}")
|
||||
print(f" Avg vitality: {result['avg_vitality']:.4f}")
|
||||
print(f" Fading (<0.3): {result['fading_count']}")
|
||||
print(f" Vibrant (>0.7): {result['vibrant_count']}")
|
||||
|
||||
|
||||
def cmd_vitality(args):
|
||||
archive = MnemosyneArchive()
|
||||
try:
|
||||
v = archive.get_vitality(args.entry_id)
|
||||
except KeyError:
|
||||
print(f"Entry not found: {args.entry_id}")
|
||||
sys.exit(1)
|
||||
print(f"[{v['entry_id'][:8]}] {v['title']}")
|
||||
print(f" Vitality: {v['vitality']:.4f}")
|
||||
print(f" Last accessed: {v['last_accessed'] or 'never'}")
|
||||
print(f" Age: {v['age_days']} days")
|
||||
|
||||
|
||||
def cmd_fading(args):
|
||||
archive = MnemosyneArchive()
|
||||
results = archive.fading(limit=args.limit)
|
||||
if not results:
|
||||
print("Archive is empty.")
|
||||
return
|
||||
for v in results:
|
||||
print(f"[{v['entry_id'][:8]}] {v['title']}")
|
||||
print(f" Vitality: {v['vitality']:.4f} | Age: {v['age_days']}d | Last: {v['last_accessed'] or 'never'}")
|
||||
print()
|
||||
|
||||
|
||||
def cmd_snapshot(args):
|
||||
archive = MnemosyneArchive()
|
||||
if args.snapshot_cmd == "create":
|
||||
result = archive.snapshot_create(label=args.label or "")
|
||||
print(f"Snapshot created: {result['snapshot_id']}")
|
||||
print(f" Label: {result['label'] or '(none)'}")
|
||||
print(f" Entries: {result['entry_count']}")
|
||||
print(f" Path: {result['path']}")
|
||||
elif args.snapshot_cmd == "list":
|
||||
snapshots = archive.snapshot_list()
|
||||
if not snapshots:
|
||||
print("No snapshots found.")
|
||||
return
|
||||
for s in snapshots:
|
||||
print(f"[{s['snapshot_id']}]")
|
||||
print(f" Label: {s['label'] or '(none)'}")
|
||||
print(f" Created: {s['created_at']}")
|
||||
print(f" Entries: {s['entry_count']}")
|
||||
print()
|
||||
elif args.snapshot_cmd == "restore":
|
||||
try:
|
||||
result = archive.snapshot_restore(args.snapshot_id)
|
||||
except FileNotFoundError as e:
|
||||
print(str(e))
|
||||
sys.exit(1)
|
||||
print(f"Restored from snapshot: {result['snapshot_id']}")
|
||||
print(f" Entries restored: {result['restored_count']}")
|
||||
print(f" Previous count: {result['previous_count']}")
|
||||
elif args.snapshot_cmd == "diff":
|
||||
try:
|
||||
diff = archive.snapshot_diff(args.snapshot_id)
|
||||
except FileNotFoundError as e:
|
||||
print(str(e))
|
||||
sys.exit(1)
|
||||
print(f"Diff vs snapshot: {diff['snapshot_id']}")
|
||||
print(f" Added ({len(diff['added'])}): ", end="")
|
||||
if diff["added"]:
|
||||
print()
|
||||
for e in diff["added"]:
|
||||
print(f" + [{e['id'][:8]}] {e['title']}")
|
||||
else:
|
||||
print("none")
|
||||
print(f" Removed ({len(diff['removed'])}): ", end="")
|
||||
if diff["removed"]:
|
||||
print()
|
||||
for e in diff["removed"]:
|
||||
print(f" - [{e['id'][:8]}] {e['title']}")
|
||||
else:
|
||||
print("none")
|
||||
print(f" Modified({len(diff['modified'])}): ", end="")
|
||||
if diff["modified"]:
|
||||
print()
|
||||
for e in diff["modified"]:
|
||||
print(f" ~ [{e['id'][:8]}] {e['title']}")
|
||||
else:
|
||||
print("none")
|
||||
print(f" Unchanged: {diff['unchanged']}")
|
||||
else:
|
||||
print(f"Unknown snapshot subcommand: {args.snapshot_cmd}")
|
||||
sys.exit(1)
|
||||
|
||||
|
||||
def cmd_resonance(args):
|
||||
archive = MnemosyneArchive()
|
||||
topic = args.topic if args.topic else None
|
||||
pairs = archive.resonance(threshold=args.threshold, limit=args.limit, topic=topic)
|
||||
if not pairs:
|
||||
print("No resonant pairs found.")
|
||||
return
|
||||
for p in pairs:
|
||||
a = p["entry_a"]
|
||||
b = p["entry_b"]
|
||||
print(f"Score: {p['score']:.4f}")
|
||||
print(f" [{a['id'][:8]}] {a['title']}")
|
||||
print(f" Topics: {', '.join(a['topics']) if a['topics'] else '(none)'}")
|
||||
print(f" [{b['id'][:8]}] {b['title']}")
|
||||
print(f" Topics: {', '.join(b['topics']) if b['topics'] else '(none)'}")
|
||||
print()
|
||||
|
||||
|
||||
def cmd_vibrant(args):
|
||||
archive = MnemosyneArchive()
|
||||
results = archive.vibrant(limit=args.limit)
|
||||
if not results:
|
||||
print("Archive is empty.")
|
||||
return
|
||||
for v in results:
|
||||
print(f"[{v['entry_id'][:8]}] {v['title']}")
|
||||
print(f" Vitality: {v['vitality']:.4f} | Age: {v['age_days']}d | Last: {v['last_accessed'] or 'never'}")
|
||||
print()
|
||||
|
||||
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(prog="mnemosyne", description="The Living Holographic Archive")
|
||||
sub = parser.add_subparsers(dest="command")
|
||||
@@ -196,6 +420,10 @@ def main():
|
||||
i.add_argument("--content", required=True)
|
||||
i.add_argument("--topics", default="", help="Comma-separated topics")
|
||||
|
||||
id_ = sub.add_parser("ingest-dir", help="Ingest a directory of files")
|
||||
id_.add_argument("path", help="Directory to ingest")
|
||||
id_.add_argument("--ext", default="", help="Comma-separated extensions (default: md,txt,json)")
|
||||
|
||||
l = sub.add_parser("link", help="Show linked entries")
|
||||
l.add_argument("entry_id", help="Entry ID (or prefix)")
|
||||
l.add_argument("-d", "--depth", type=int, default=1)
|
||||
@@ -233,15 +461,67 @@ def main():
|
||||
rt.add_argument("entry_id", help="Entry ID")
|
||||
rt.add_argument("tags", help="Comma-separated new tag list")
|
||||
|
||||
tl = sub.add_parser("timeline", help="Show entries within an ISO date range")
|
||||
tl.add_argument("start", help="Start datetime (ISO format, e.g. 2024-01-01 or 2024-01-01T00:00:00Z)")
|
||||
tl.add_argument("end", help="End datetime (ISO format)")
|
||||
|
||||
nb = sub.add_parser("neighbors", help="Show entries temporally near a given entry")
|
||||
nb.add_argument("entry_id", help="Anchor entry ID")
|
||||
nb.add_argument("--days", type=int, default=7, help="Window in days (default: 7)")
|
||||
|
||||
|
||||
pa = sub.add_parser("path", help="Find shortest path between two memories")
|
||||
pa.add_argument("start", help="Starting entry ID")
|
||||
pa.add_argument("end", help="Target entry ID")
|
||||
pa.add_argument("--archive", default=None, help="Archive path")
|
||||
|
||||
co = sub.add_parser("consolidate", help="Merge duplicate/near-duplicate entries")
|
||||
co.add_argument("--dry-run", action="store_true", help="Show what would be merged without applying")
|
||||
co.add_argument("--threshold", type=float, default=0.9, help="Similarity threshold (default: 0.9)")
|
||||
|
||||
|
||||
tc = sub.add_parser("touch", help="Boost an entry's vitality by accessing it")
|
||||
tc.add_argument("entry_id", help="Entry ID to touch")
|
||||
|
||||
dc = sub.add_parser("decay", help="Apply time-based decay to all entries")
|
||||
|
||||
vy = sub.add_parser("vitality", help="Show an entry's vitality status")
|
||||
vy.add_argument("entry_id", help="Entry ID to check")
|
||||
|
||||
fg = sub.add_parser("fading", help="Show most neglected entries (lowest vitality)")
|
||||
fg.add_argument("-n", "--limit", type=int, default=10, help="Max entries to show")
|
||||
|
||||
vb = sub.add_parser("vibrant", help="Show most alive entries (highest vitality)")
|
||||
vb.add_argument("-n", "--limit", type=int, default=10, help="Max entries to show")
|
||||
|
||||
rs = sub.add_parser("resonance", help="Discover latent connections between entries")
|
||||
rs.add_argument("-t", "--threshold", type=float, default=0.3, help="Minimum similarity score (default: 0.3)")
|
||||
rs.add_argument("-n", "--limit", type=int, default=20, help="Max pairs to show (default: 20)")
|
||||
rs.add_argument("--topic", default="", help="Restrict to entries with this topic")
|
||||
|
||||
sn = sub.add_parser("snapshot", help="Point-in-time backup and restore")
|
||||
sn_sub = sn.add_subparsers(dest="snapshot_cmd")
|
||||
sn_create = sn_sub.add_parser("create", help="Create a new snapshot")
|
||||
sn_create.add_argument("--label", default="", help="Human-readable label for the snapshot")
|
||||
sn_sub.add_parser("list", help="List available snapshots")
|
||||
sn_restore = sn_sub.add_parser("restore", help="Restore archive from a snapshot")
|
||||
sn_restore.add_argument("snapshot_id", help="Snapshot ID to restore")
|
||||
sn_diff = sn_sub.add_parser("diff", help="Show what changed since a snapshot")
|
||||
sn_diff.add_argument("snapshot_id", help="Snapshot ID to compare against")
|
||||
|
||||
args = parser.parse_args()
|
||||
if not args.command:
|
||||
parser.print_help()
|
||||
sys.exit(1)
|
||||
if args.command == "snapshot" and not args.snapshot_cmd:
|
||||
sn.print_help()
|
||||
sys.exit(1)
|
||||
|
||||
dispatch = {
|
||||
"stats": cmd_stats,
|
||||
"search": cmd_search,
|
||||
"ingest": cmd_ingest,
|
||||
"ingest-dir": cmd_ingest_dir,
|
||||
"link": cmd_link,
|
||||
"topics": cmd_topics,
|
||||
"remove": cmd_remove,
|
||||
@@ -253,6 +533,17 @@ def main():
|
||||
"tag": cmd_tag,
|
||||
"untag": cmd_untag,
|
||||
"retag": cmd_retag,
|
||||
"timeline": cmd_timeline,
|
||||
"neighbors": cmd_neighbors,
|
||||
"consolidate": cmd_consolidate,
|
||||
"path": cmd_path,
|
||||
"touch": cmd_touch,
|
||||
"decay": cmd_decay,
|
||||
"vitality": cmd_vitality,
|
||||
"fading": cmd_fading,
|
||||
"vibrant": cmd_vibrant,
|
||||
"resonance": cmd_resonance,
|
||||
"snapshot": cmd_snapshot,
|
||||
}
|
||||
dispatch[args.command](args)
|
||||
|
||||
|
||||
170
nexus/mnemosyne/embeddings.py
Normal file
170
nexus/mnemosyne/embeddings.py
Normal file
@@ -0,0 +1,170 @@
|
||||
"""Pluggable embedding backends for Mnemosyne semantic search.
|
||||
|
||||
Provides an abstract EmbeddingBackend interface and concrete implementations:
|
||||
- OllamaEmbeddingBackend: local models via Ollama (sovereign, no cloud)
|
||||
- TfidfEmbeddingBackend: pure-Python TF-IDF fallback (no dependencies)
|
||||
|
||||
Usage:
|
||||
from nexus.mnemosyne.embeddings import get_embedding_backend
|
||||
backend = get_embedding_backend() # auto-detects best available
|
||||
vec = backend.embed("hello world")
|
||||
score = backend.similarity(vec_a, vec_b)
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
import abc, json, math, os, re, urllib.request
|
||||
from typing import Optional
|
||||
|
||||
|
||||
class EmbeddingBackend(abc.ABC):
|
||||
"""Abstract interface for embedding-based similarity."""
|
||||
|
||||
@abc.abstractmethod
|
||||
def embed(self, text: str) -> list[float]:
|
||||
"""Return an embedding vector for the given text."""
|
||||
|
||||
@abc.abstractmethod
|
||||
def similarity(self, a: list[float], b: list[float]) -> float:
|
||||
"""Return cosine similarity between two vectors, in [0, 1]."""
|
||||
|
||||
@property
|
||||
def name(self) -> str:
|
||||
return self.__class__.__name__
|
||||
|
||||
@property
|
||||
def dimension(self) -> int:
|
||||
return 0
|
||||
|
||||
|
||||
def cosine_similarity(a: list[float], b: list[float]) -> float:
|
||||
"""Cosine similarity between two vectors."""
|
||||
if len(a) != len(b):
|
||||
raise ValueError(f"Vector dimension mismatch: {len(a)} vs {len(b)}")
|
||||
dot = sum(x * y for x, y in zip(a, b))
|
||||
norm_a = math.sqrt(sum(x * x for x in a))
|
||||
norm_b = math.sqrt(sum(x * x for x in b))
|
||||
if norm_a == 0 or norm_b == 0:
|
||||
return 0.0
|
||||
return dot / (norm_a * norm_b)
|
||||
|
||||
|
||||
class OllamaEmbeddingBackend(EmbeddingBackend):
|
||||
"""Embedding backend using a local Ollama instance.
|
||||
Default model: nomic-embed-text (768 dims)."""
|
||||
|
||||
def __init__(self, base_url: str | None = None, model: str | None = None):
|
||||
self.base_url = base_url or os.environ.get("OLLAMA_URL", "http://localhost:11434")
|
||||
self.model = model or os.environ.get("MNEMOSYNE_EMBED_MODEL", "nomic-embed-text")
|
||||
self._dim: int = 0
|
||||
self._available: bool | None = None
|
||||
|
||||
def _check_available(self) -> bool:
|
||||
if self._available is not None:
|
||||
return self._available
|
||||
try:
|
||||
req = urllib.request.Request(f"{self.base_url}/api/tags", method="GET")
|
||||
resp = urllib.request.urlopen(req, timeout=3)
|
||||
tags = json.loads(resp.read())
|
||||
models = [m["name"].split(":")[0] for m in tags.get("models", [])]
|
||||
self._available = any(self.model in m for m in models)
|
||||
except Exception:
|
||||
self._available = False
|
||||
return self._available
|
||||
|
||||
@property
|
||||
def name(self) -> str:
|
||||
return f"Ollama({self.model})"
|
||||
|
||||
@property
|
||||
def dimension(self) -> int:
|
||||
return self._dim
|
||||
|
||||
def embed(self, text: str) -> list[float]:
|
||||
if not self._check_available():
|
||||
raise RuntimeError(f"Ollama not available or model {self.model} not found")
|
||||
data = json.dumps({"model": self.model, "prompt": text}).encode()
|
||||
req = urllib.request.Request(
|
||||
f"{self.base_url}/api/embeddings", data=data,
|
||||
headers={"Content-Type": "application/json"}, method="POST")
|
||||
resp = urllib.request.urlopen(req, timeout=30)
|
||||
result = json.loads(resp.read())
|
||||
vec = result.get("embedding", [])
|
||||
if vec:
|
||||
self._dim = len(vec)
|
||||
return vec
|
||||
|
||||
def similarity(self, a: list[float], b: list[float]) -> float:
|
||||
raw = cosine_similarity(a, b)
|
||||
return (raw + 1.0) / 2.0
|
||||
|
||||
|
||||
class TfidfEmbeddingBackend(EmbeddingBackend):
|
||||
"""Pure-Python TF-IDF embedding. No dependencies. Always available."""
|
||||
|
||||
def __init__(self):
|
||||
self._vocab: dict[str, int] = {}
|
||||
self._idf: dict[str, float] = {}
|
||||
self._doc_count: int = 0
|
||||
self._doc_freq: dict[str, int] = {}
|
||||
|
||||
@property
|
||||
def name(self) -> str:
|
||||
return "TF-IDF (local)"
|
||||
|
||||
@property
|
||||
def dimension(self) -> int:
|
||||
return len(self._vocab)
|
||||
|
||||
@staticmethod
|
||||
def _tokenize(text: str) -> list[str]:
|
||||
return [t for t in re.findall(r"\w+", text.lower()) if len(t) > 2]
|
||||
|
||||
def _update_idf(self, tokens: list[str]):
|
||||
self._doc_count += 1
|
||||
for t in set(tokens):
|
||||
self._doc_freq[t] = self._doc_freq.get(t, 0) + 1
|
||||
for t, df in self._doc_freq.items():
|
||||
self._idf[t] = math.log((self._doc_count + 1) / (df + 1)) + 1.0
|
||||
|
||||
def embed(self, text: str) -> list[float]:
|
||||
tokens = self._tokenize(text)
|
||||
if not tokens:
|
||||
return []
|
||||
for t in tokens:
|
||||
if t not in self._vocab:
|
||||
self._vocab[t] = len(self._vocab)
|
||||
self._update_idf(tokens)
|
||||
dim = len(self._vocab)
|
||||
vec = [0.0] * dim
|
||||
tf = {}
|
||||
for t in tokens:
|
||||
tf[t] = tf.get(t, 0) + 1
|
||||
for t, count in tf.items():
|
||||
vec[self._vocab[t]] = (count / len(tokens)) * self._idf.get(t, 1.0)
|
||||
norm = math.sqrt(sum(v * v for v in vec))
|
||||
if norm > 0:
|
||||
vec = [v / norm for v in vec]
|
||||
return vec
|
||||
|
||||
def similarity(self, a: list[float], b: list[float]) -> float:
|
||||
if len(a) != len(b):
|
||||
mx = max(len(a), len(b))
|
||||
a = a + [0.0] * (mx - len(a))
|
||||
b = b + [0.0] * (mx - len(b))
|
||||
return max(0.0, cosine_similarity(a, b))
|
||||
|
||||
|
||||
def get_embedding_backend(prefer: str | None = None, ollama_url: str | None = None,
|
||||
model: str | None = None) -> EmbeddingBackend:
|
||||
"""Auto-detect best available embedding backend. Priority: Ollama > TF-IDF."""
|
||||
env_pref = os.environ.get("MNEMOSYNE_EMBED_BACKEND")
|
||||
effective = prefer or env_pref
|
||||
if effective == "tfidf":
|
||||
return TfidfEmbeddingBackend()
|
||||
if effective in (None, "ollama"):
|
||||
ollama = OllamaEmbeddingBackend(base_url=ollama_url, model=model)
|
||||
if ollama._check_available():
|
||||
return ollama
|
||||
if effective == "ollama":
|
||||
raise RuntimeError("Ollama backend requested but not available")
|
||||
return TfidfEmbeddingBackend()
|
||||
@@ -13,6 +13,12 @@ from typing import Optional
|
||||
import uuid
|
||||
|
||||
|
||||
def _compute_content_hash(title: str, content: str) -> str:
|
||||
"""Compute SHA-256 of title+content for deduplication."""
|
||||
raw = f"{title}\x00{content}".encode("utf-8")
|
||||
return hashlib.sha256(raw).hexdigest()
|
||||
|
||||
|
||||
@dataclass
|
||||
class ArchiveEntry:
|
||||
"""A single node in the Mnemosyne holographic archive."""
|
||||
@@ -25,18 +31,15 @@ class ArchiveEntry:
|
||||
topics: list[str] = field(default_factory=list)
|
||||
metadata: dict = field(default_factory=dict)
|
||||
created_at: str = field(default_factory=lambda: datetime.now(timezone.utc).isoformat())
|
||||
updated_at: str = field(default_factory=lambda: datetime.now(timezone.utc).isoformat())
|
||||
updated_at: Optional[str] = None # Set on mutation; None means same as created_at
|
||||
links: list[str] = field(default_factory=list) # IDs of related entries
|
||||
content_hash: Optional[str] = None # SHA-256 of title+content for dedup
|
||||
vitality: float = 1.0 # 0.0 (dead) to 1.0 (fully alive)
|
||||
last_accessed: Optional[str] = None # ISO datetime of last access; None = never accessed
|
||||
|
||||
@property
|
||||
def content_hash(self) -> str:
|
||||
"""SHA-256 hash of title + content for dedup detection."""
|
||||
raw = f"{self.title}\x00{self.content}".encode()
|
||||
return hashlib.sha256(raw).hexdigest()
|
||||
|
||||
def touch(self):
|
||||
"""Bump updated_at to now."""
|
||||
self.updated_at = datetime.now(timezone.utc).isoformat()
|
||||
def __post_init__(self):
|
||||
if self.content_hash is None:
|
||||
self.content_hash = _compute_content_hash(self.title, self.content)
|
||||
|
||||
def to_dict(self) -> dict:
|
||||
return {
|
||||
@@ -51,13 +54,10 @@ class ArchiveEntry:
|
||||
"updated_at": self.updated_at,
|
||||
"links": self.links,
|
||||
"content_hash": self.content_hash,
|
||||
"vitality": self.vitality,
|
||||
"last_accessed": self.last_accessed,
|
||||
}
|
||||
|
||||
@classmethod
|
||||
def from_dict(cls, data: dict) -> ArchiveEntry:
|
||||
# Strip non-field keys (like content_hash which is computed)
|
||||
filtered = {k: v for k, v in data.items() if k in cls.__dataclass_fields__}
|
||||
# Backfill updated_at for legacy entries that lack it
|
||||
if "updated_at" not in filtered:
|
||||
filtered["updated_at"] = filtered.get("created_at", datetime.now(timezone.utc).isoformat())
|
||||
return cls(**filtered)
|
||||
return cls(**{k: v for k, v in data.items() if k in cls.__dataclass_fields__})
|
||||
|
||||
@@ -1,15 +1,135 @@
|
||||
"""Ingestion pipeline — feeds data into the archive.
|
||||
|
||||
Supports ingesting from MemPalace, raw events, and manual entries.
|
||||
Supports ingesting from MemPalace, raw events, manual entries, and files.
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import Optional
|
||||
import re
|
||||
from pathlib import Path
|
||||
from typing import Optional, Union
|
||||
|
||||
from nexus.mnemosyne.archive import MnemosyneArchive
|
||||
from nexus.mnemosyne.entry import ArchiveEntry
|
||||
|
||||
_DEFAULT_EXTENSIONS = [".md", ".txt", ".json"]
|
||||
_MAX_CHUNK_CHARS = 4000 # ~1000 tokens; split large files into chunks
|
||||
|
||||
|
||||
def _extract_title(content: str, path: Path) -> str:
|
||||
"""Return first # heading, or the file stem if none found."""
|
||||
for line in content.splitlines():
|
||||
stripped = line.strip()
|
||||
if stripped.startswith("# "):
|
||||
return stripped[2:].strip()
|
||||
return path.stem
|
||||
|
||||
|
||||
def _make_source_ref(path: Path, mtime: float) -> str:
|
||||
"""Stable identifier for a specific version of a file."""
|
||||
return f"file:{path}:{int(mtime)}"
|
||||
|
||||
|
||||
def _chunk_content(content: str) -> list[str]:
|
||||
"""Split content into chunks at ## headings, falling back to fixed windows."""
|
||||
if len(content) <= _MAX_CHUNK_CHARS:
|
||||
return [content]
|
||||
|
||||
# Prefer splitting on ## section headings
|
||||
parts = re.split(r"\n(?=## )", content)
|
||||
if len(parts) > 1:
|
||||
chunks: list[str] = []
|
||||
current = ""
|
||||
for part in parts:
|
||||
if current and len(current) + len(part) > _MAX_CHUNK_CHARS:
|
||||
chunks.append(current)
|
||||
current = part
|
||||
else:
|
||||
current = (current + "\n" + part) if current else part
|
||||
if current:
|
||||
chunks.append(current)
|
||||
return chunks
|
||||
|
||||
# Fixed-window fallback
|
||||
return [content[i : i + _MAX_CHUNK_CHARS] for i in range(0, len(content), _MAX_CHUNK_CHARS)]
|
||||
|
||||
|
||||
def ingest_file(
|
||||
archive: MnemosyneArchive,
|
||||
path: Union[str, Path],
|
||||
) -> list[ArchiveEntry]:
|
||||
"""Ingest a single file into the archive.
|
||||
|
||||
- Title is taken from the first ``# heading`` or the filename stem.
|
||||
- Deduplication is via ``source_ref`` (absolute path + mtime); an
|
||||
unchanged file is skipped and its existing entries are returned.
|
||||
- Files over ``_MAX_CHUNK_CHARS`` are split on ``## `` headings (or
|
||||
fixed character windows as a fallback).
|
||||
|
||||
Returns a list of ArchiveEntry objects (one per chunk).
|
||||
"""
|
||||
path = Path(path).resolve()
|
||||
mtime = path.stat().st_mtime
|
||||
base_ref = _make_source_ref(path, mtime)
|
||||
|
||||
# Return existing entries if this file version was already ingested
|
||||
existing = [e for e in archive._entries.values() if e.source_ref and e.source_ref.startswith(base_ref)]
|
||||
if existing:
|
||||
return existing
|
||||
|
||||
content = path.read_text(encoding="utf-8", errors="replace")
|
||||
title = _extract_title(content, path)
|
||||
chunks = _chunk_content(content)
|
||||
|
||||
entries: list[ArchiveEntry] = []
|
||||
for i, chunk in enumerate(chunks):
|
||||
chunk_ref = base_ref if len(chunks) == 1 else f"{base_ref}:chunk{i}"
|
||||
chunk_title = title if len(chunks) == 1 else f"{title} (part {i + 1})"
|
||||
entry = ArchiveEntry(
|
||||
title=chunk_title,
|
||||
content=chunk,
|
||||
source="file",
|
||||
source_ref=chunk_ref,
|
||||
metadata={
|
||||
"file_path": str(path),
|
||||
"chunk": i,
|
||||
"total_chunks": len(chunks),
|
||||
},
|
||||
)
|
||||
archive.add(entry)
|
||||
entries.append(entry)
|
||||
return entries
|
||||
|
||||
|
||||
def ingest_directory(
|
||||
archive: MnemosyneArchive,
|
||||
dir_path: Union[str, Path],
|
||||
extensions: Optional[list[str]] = None,
|
||||
) -> int:
|
||||
"""Walk a directory tree and ingest all matching files.
|
||||
|
||||
``extensions`` defaults to ``[".md", ".txt", ".json"]``.
|
||||
Values may be given with or without a leading dot.
|
||||
|
||||
Returns the count of new archive entries created.
|
||||
"""
|
||||
dir_path = Path(dir_path).resolve()
|
||||
if extensions is None:
|
||||
exts = _DEFAULT_EXTENSIONS
|
||||
else:
|
||||
exts = [e if e.startswith(".") else f".{e}" for e in extensions]
|
||||
|
||||
added = 0
|
||||
for file_path in sorted(dir_path.rglob("*")):
|
||||
if not file_path.is_file():
|
||||
continue
|
||||
if file_path.suffix.lower() not in exts:
|
||||
continue
|
||||
before = archive.count
|
||||
ingest_file(archive, file_path)
|
||||
added += archive.count - before
|
||||
return added
|
||||
|
||||
|
||||
def ingest_from_mempalace(
|
||||
archive: MnemosyneArchive,
|
||||
|
||||
@@ -2,31 +2,63 @@
|
||||
|
||||
Computes semantic similarity between archive entries and creates
|
||||
bidirectional links, forming the holographic graph structure.
|
||||
|
||||
Supports pluggable embedding backends for true semantic search.
|
||||
Falls back to Jaccard token similarity when no backend is available.
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
from typing import Optional
|
||||
from typing import Optional, TYPE_CHECKING
|
||||
|
||||
from nexus.mnemosyne.entry import ArchiveEntry
|
||||
|
||||
if TYPE_CHECKING:
|
||||
from nexus.mnemosyne.embeddings import EmbeddingBackend
|
||||
|
||||
|
||||
class HolographicLinker:
|
||||
"""Links archive entries via semantic similarity.
|
||||
|
||||
Phase 1 uses simple keyword overlap as the similarity metric.
|
||||
Phase 2 will integrate ChromaDB embeddings from MemPalace.
|
||||
With an embedding backend: cosine similarity on vectors.
|
||||
Without: Jaccard similarity on token sets (legacy fallback).
|
||||
"""
|
||||
|
||||
def __init__(self, similarity_threshold: float = 0.15):
|
||||
def __init__(
|
||||
self,
|
||||
similarity_threshold: float = 0.15,
|
||||
embedding_backend: Optional["EmbeddingBackend"] = None,
|
||||
):
|
||||
self.threshold = similarity_threshold
|
||||
self._backend = embedding_backend
|
||||
self._embed_cache: dict[str, list[float]] = {}
|
||||
|
||||
@property
|
||||
def using_embeddings(self) -> bool:
|
||||
return self._backend is not None
|
||||
|
||||
def _get_embedding(self, entry: ArchiveEntry) -> list[float]:
|
||||
"""Get or compute cached embedding for an entry."""
|
||||
if entry.id in self._embed_cache:
|
||||
return self._embed_cache[entry.id]
|
||||
text = f"{entry.title} {entry.content}"
|
||||
vec = self._backend.embed(text) if self._backend else []
|
||||
if vec:
|
||||
self._embed_cache[entry.id] = vec
|
||||
return vec
|
||||
|
||||
def compute_similarity(self, a: ArchiveEntry, b: ArchiveEntry) -> float:
|
||||
"""Compute similarity score between two entries.
|
||||
|
||||
Returns float in [0, 1]. Phase 1: Jaccard similarity on
|
||||
combined title+content tokens. Phase 2: cosine similarity
|
||||
on ChromaDB embeddings.
|
||||
Returns float in [0, 1]. Uses embedding cosine similarity if
|
||||
a backend is configured, otherwise falls back to Jaccard.
|
||||
"""
|
||||
if self._backend:
|
||||
vec_a = self._get_embedding(a)
|
||||
vec_b = self._get_embedding(b)
|
||||
if vec_a and vec_b:
|
||||
return self._backend.similarity(vec_a, vec_b)
|
||||
# Fallback: Jaccard on tokens
|
||||
tokens_a = self._tokenize(f"{a.title} {a.content}")
|
||||
tokens_b = self._tokenize(f"{b.title} {b.content}")
|
||||
if not tokens_a or not tokens_b:
|
||||
@@ -35,11 +67,10 @@ class HolographicLinker:
|
||||
union = tokens_a | tokens_b
|
||||
return len(intersection) / len(union)
|
||||
|
||||
def find_links(self, entry: ArchiveEntry, candidates: list[ArchiveEntry]) -> list[tuple[str, float]]:
|
||||
"""Find entries worth linking to.
|
||||
|
||||
Returns list of (entry_id, similarity_score) tuples above threshold.
|
||||
"""
|
||||
def find_links(
|
||||
self, entry: ArchiveEntry, candidates: list[ArchiveEntry]
|
||||
) -> list[tuple[str, float]]:
|
||||
"""Find entries worth linking to. Returns (entry_id, score) tuples."""
|
||||
results = []
|
||||
for candidate in candidates:
|
||||
if candidate.id == entry.id:
|
||||
@@ -58,16 +89,18 @@ class HolographicLinker:
|
||||
if eid not in entry.links:
|
||||
entry.links.append(eid)
|
||||
new_links += 1
|
||||
# Bidirectional
|
||||
for c in candidates:
|
||||
if c.id == eid and entry.id not in c.links:
|
||||
c.links.append(entry.id)
|
||||
return new_links
|
||||
|
||||
def clear_cache(self):
|
||||
"""Clear embedding cache (call after bulk entry changes)."""
|
||||
self._embed_cache.clear()
|
||||
|
||||
@staticmethod
|
||||
def _tokenize(text: str) -> set[str]:
|
||||
"""Simple whitespace + punctuation tokenizer."""
|
||||
import re
|
||||
tokens = set(re.findall(r"\w+", text.lower()))
|
||||
# Remove very short tokens
|
||||
return {t for t in tokens if len(t) > 2}
|
||||
|
||||
14
nexus/mnemosyne/reasoner.py
Normal file
14
nexus/mnemosyne/reasoner.py
Normal file
@@ -0,0 +1,14 @@
|
||||
|
||||
class Reasoner:
|
||||
def __init__(self, rules):
|
||||
self.rules = rules
|
||||
def evaluate(self, entries):
|
||||
return [r['action'] for r in self.rules if self._check(r['condition'], entries)]
|
||||
def _check(self, cond, entries):
|
||||
if cond.startswith('count'):
|
||||
# e.g. count(type=anomaly)>3
|
||||
p = cond.replace('count(', '').split(')')
|
||||
key, val = p[0].split('=')
|
||||
count = sum(1 for e in entries if e.get(key) == val)
|
||||
return eval(f"{count}{p[1]}")
|
||||
return False
|
||||
6
nexus/mnemosyne/rules.json
Normal file
6
nexus/mnemosyne/rules.json
Normal file
@@ -0,0 +1,6 @@
|
||||
[
|
||||
{
|
||||
"condition": "count(type=anomaly)>3",
|
||||
"action": "alert"
|
||||
}
|
||||
]
|
||||
2
nexus/mnemosyne/snapshot.py
Normal file
2
nexus/mnemosyne/snapshot.py
Normal file
@@ -0,0 +1,2 @@
|
||||
import json
|
||||
# Snapshot logic
|
||||
Binary file not shown.
Binary file not shown.
Binary file not shown.
@@ -2,6 +2,7 @@
|
||||
|
||||
import json
|
||||
import tempfile
|
||||
from datetime import datetime, timezone, timedelta
|
||||
from pathlib import Path
|
||||
|
||||
from nexus.mnemosyne.entry import ArchiveEntry
|
||||
@@ -493,66 +494,112 @@ def test_tag_persistence_across_reload():
|
||||
assert "alpha" not in fresh.topics
|
||||
|
||||
|
||||
# --- Entry update + dedup tests ---
|
||||
# --- content_hash and updated_at field tests ---
|
||||
|
||||
def test_content_hash_deterministic():
|
||||
e1 = ArchiveEntry(title="Test", content="Hello")
|
||||
e2 = ArchiveEntry(title="Test", content="Hello")
|
||||
def test_entry_has_content_hash():
|
||||
e = ArchiveEntry(title="Hello", content="world")
|
||||
assert e.content_hash is not None
|
||||
assert len(e.content_hash) == 64 # SHA-256 hex
|
||||
|
||||
|
||||
def test_entry_content_hash_deterministic():
|
||||
e1 = ArchiveEntry(title="Hello", content="world")
|
||||
e2 = ArchiveEntry(title="Hello", content="world")
|
||||
assert e1.content_hash == e2.content_hash
|
||||
|
||||
|
||||
def test_content_hash_differs_on_change():
|
||||
e = ArchiveEntry(title="Test", content="Hello")
|
||||
h1 = e.content_hash
|
||||
e.content = "World"
|
||||
assert e.content_hash != h1
|
||||
def test_entry_content_hash_differs_on_different_content():
|
||||
e1 = ArchiveEntry(title="Hello", content="world")
|
||||
e2 = ArchiveEntry(title="Hello", content="different")
|
||||
assert e1.content_hash != e2.content_hash
|
||||
|
||||
|
||||
def test_updated_at_set_on_creation():
|
||||
def test_entry_updated_at_defaults_none():
|
||||
e = ArchiveEntry(title="T", content="c")
|
||||
assert e.updated_at is not None
|
||||
assert e.updated_at >= e.created_at
|
||||
assert e.updated_at is None
|
||||
|
||||
|
||||
def test_touch_updates_timestamp():
|
||||
import time
|
||||
def test_entry_roundtrip_includes_new_fields():
|
||||
e = ArchiveEntry(title="T", content="c")
|
||||
before = e.updated_at
|
||||
time.sleep(0.01)
|
||||
e.touch()
|
||||
assert e.updated_at >= before
|
||||
d = e.to_dict()
|
||||
assert "content_hash" in d
|
||||
assert "updated_at" in d
|
||||
e2 = ArchiveEntry.from_dict(d)
|
||||
assert e2.content_hash == e.content_hash
|
||||
assert e2.updated_at == e.updated_at
|
||||
|
||||
|
||||
# --- content deduplication tests ---
|
||||
|
||||
def test_add_deduplication_same_content():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e1 = ingest_event(archive, title="Dup", content="Same content here")
|
||||
e2 = ingest_event(archive, title="Dup", content="Same content here")
|
||||
# Should NOT have created a second entry
|
||||
assert archive.count == 1
|
||||
assert e1.id == e2.id
|
||||
|
||||
|
||||
def test_add_deduplication_different_content():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
ingest_event(archive, title="A", content="Content one")
|
||||
ingest_event(archive, title="B", content="Content two")
|
||||
assert archive.count == 2
|
||||
|
||||
|
||||
def test_find_duplicate_returns_existing():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e1 = ingest_event(archive, title="Dup", content="Same content here")
|
||||
probe = ArchiveEntry(title="Dup", content="Same content here")
|
||||
dup = archive.find_duplicate(probe)
|
||||
assert dup is not None
|
||||
assert dup.id == e1.id
|
||||
|
||||
|
||||
def test_find_duplicate_returns_none_for_unique():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
ingest_event(archive, title="A", content="Some content")
|
||||
probe = ArchiveEntry(title="B", content="Totally different content")
|
||||
assert archive.find_duplicate(probe) is None
|
||||
|
||||
|
||||
def test_find_duplicate_empty_archive():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
probe = ArchiveEntry(title="X", content="y")
|
||||
assert archive.find_duplicate(probe) is None
|
||||
|
||||
|
||||
# --- update_entry tests ---
|
||||
|
||||
def test_update_entry_title():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e = ingest_event(archive, title="Old", content="content", topics=["x"])
|
||||
old_hash = e.content_hash
|
||||
updated = archive.update_entry(e.id, title="New Title")
|
||||
assert updated.title == "New Title"
|
||||
assert updated.content == "content"
|
||||
assert updated.updated_at >= e.created_at
|
||||
# Content unchanged, so hash should be same (only title changed)
|
||||
assert updated.content_hash != old_hash # title is in hash
|
||||
e = ingest_event(archive, title="Old title", content="Some content")
|
||||
archive.update_entry(e.id, title="New title")
|
||||
fresh = archive.get(e.id)
|
||||
assert fresh.title == "New title"
|
||||
assert fresh.content == "Some content"
|
||||
|
||||
|
||||
def test_update_entry_content_relinks():
|
||||
def test_update_entry_content():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e1 = ingest_event(archive, title="Python", content="Python programming language")
|
||||
e2 = ingest_event(archive, title="Java", content="Java programming language")
|
||||
# e1 and e2 should be linked via shared tokens
|
||||
assert e2.id in e1.links or e1.id in e2.links
|
||||
|
||||
# Update e1 to completely different content
|
||||
archive.update_entry(e1.id, content="Cooking recipes for dinner")
|
||||
e1_fresh = archive.get(e1.id)
|
||||
e2_fresh = archive.get(e2.id)
|
||||
# e1 should have been re-linked (likely unlinked from e2 now)
|
||||
# e2 should no longer reference e1
|
||||
assert e1_fresh.content == "Cooking recipes for dinner"
|
||||
e = ingest_event(archive, title="T", content="Old content")
|
||||
archive.update_entry(e.id, content="New content")
|
||||
fresh = archive.get(e.id)
|
||||
assert fresh.content == "New content"
|
||||
|
||||
|
||||
def test_update_entry_metadata():
|
||||
@@ -562,7 +609,29 @@ def test_update_entry_metadata():
|
||||
e = ingest_event(archive, title="T", content="c")
|
||||
archive.update_entry(e.id, metadata={"key": "value"})
|
||||
fresh = archive.get(e.id)
|
||||
assert fresh.metadata == {"key": "value"}
|
||||
assert fresh.metadata["key"] == "value"
|
||||
|
||||
|
||||
def test_update_entry_bumps_updated_at():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e = ingest_event(archive, title="T", content="c")
|
||||
assert e.updated_at is None
|
||||
archive.update_entry(e.id, title="Updated")
|
||||
fresh = archive.get(e.id)
|
||||
assert fresh.updated_at is not None
|
||||
|
||||
|
||||
def test_update_entry_refreshes_content_hash():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e = ingest_event(archive, title="T", content="Original content")
|
||||
old_hash = e.content_hash
|
||||
archive.update_entry(e.id, content="Completely new content")
|
||||
fresh = archive.get(e.id)
|
||||
assert fresh.content_hash != old_hash
|
||||
|
||||
|
||||
def test_update_entry_missing_raises():
|
||||
@@ -570,113 +639,217 @@ def test_update_entry_missing_raises():
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
try:
|
||||
archive.update_entry("nonexistent", title="X")
|
||||
archive.update_entry("nonexistent-id", title="X")
|
||||
assert False, "Expected KeyError"
|
||||
except KeyError:
|
||||
pass
|
||||
|
||||
|
||||
def test_update_entry_no_change_no_relink():
|
||||
def test_update_entry_persists_across_reload():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
a1 = MnemosyneArchive(archive_path=path)
|
||||
e = ingest_event(a1, title="Before", content="Before content")
|
||||
a1.update_entry(e.id, title="After", content="After content")
|
||||
|
||||
a2 = MnemosyneArchive(archive_path=path)
|
||||
fresh = a2.get(e.id)
|
||||
assert fresh.title == "After"
|
||||
assert fresh.content == "After content"
|
||||
assert fresh.updated_at is not None
|
||||
|
||||
|
||||
def test_update_entry_no_change_no_crash():
|
||||
"""Calling update_entry with all None args should not fail."""
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e = ingest_event(archive, title="T", content="c", topics=["x"])
|
||||
orig_links = list(e.links)
|
||||
# Update only metadata (no content change)
|
||||
archive.update_entry(e.id, metadata={"k": "v"})
|
||||
fresh = archive.get(e.id)
|
||||
assert fresh.links == orig_links
|
||||
e = ingest_event(archive, title="T", content="c")
|
||||
result = archive.update_entry(e.id)
|
||||
assert result.title == "T"
|
||||
|
||||
|
||||
def test_find_by_hash():
|
||||
# --- by_date_range tests ---
|
||||
|
||||
def _make_entry_at(archive: MnemosyneArchive, title: str, dt: datetime) -> ArchiveEntry:
|
||||
"""Helper: ingest an entry and backdate its created_at."""
|
||||
e = ingest_event(archive, title=title, content=title)
|
||||
e.created_at = dt.isoformat()
|
||||
archive._save()
|
||||
return e
|
||||
|
||||
|
||||
def test_by_date_range_empty_archive():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e = ingest_event(archive, title="Unique", content="Unique content xyz")
|
||||
found = archive.find_by_hash(e.content_hash)
|
||||
assert found is not None
|
||||
assert found.id == e.id
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
results = archive.by_date_range("2024-01-01", "2024-12-31")
|
||||
assert results == []
|
||||
|
||||
|
||||
def test_find_by_hash_miss():
|
||||
def test_by_date_range_returns_matching_entries():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
found = archive.find_by_hash("nonexistent-hash")
|
||||
assert found is None
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
jan = datetime(2024, 1, 15, tzinfo=timezone.utc)
|
||||
mar = datetime(2024, 3, 10, tzinfo=timezone.utc)
|
||||
jun = datetime(2024, 6, 1, tzinfo=timezone.utc)
|
||||
e1 = _make_entry_at(archive, "Jan entry", jan)
|
||||
e2 = _make_entry_at(archive, "Mar entry", mar)
|
||||
e3 = _make_entry_at(archive, "Jun entry", jun)
|
||||
|
||||
results = archive.by_date_range("2024-01-01", "2024-04-01")
|
||||
ids = {e.id for e in results}
|
||||
assert e1.id in ids
|
||||
assert e2.id in ids
|
||||
assert e3.id not in ids
|
||||
|
||||
|
||||
def test_find_duplicates():
|
||||
def test_by_date_range_boundary_inclusive():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e1 = ingest_event(archive, title="Same", content="Duplicate content")
|
||||
# Manually add a second entry with identical title+content
|
||||
e2 = ArchiveEntry(title="Same", content="Duplicate content", source="manual")
|
||||
archive._entries[e2.id] = e2
|
||||
archive._save()
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
exact = datetime(2024, 3, 1, tzinfo=timezone.utc)
|
||||
e = _make_entry_at(archive, "Exact boundary", exact)
|
||||
|
||||
dups = archive.find_duplicates()
|
||||
assert len(dups) == 1
|
||||
assert len(dups[0]) == 2
|
||||
dup_ids = {d.id for d in dups[0]}
|
||||
assert e1.id in dup_ids
|
||||
assert e2.id in dup_ids
|
||||
results = archive.by_date_range("2024-03-01T00:00:00+00:00", "2024-03-01T00:00:00+00:00")
|
||||
assert len(results) == 1
|
||||
assert results[0].id == e.id
|
||||
|
||||
|
||||
def test_find_duplicates_none():
|
||||
def test_by_date_range_no_results():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
ingest_event(archive, title="A", content="unique a")
|
||||
ingest_event(archive, title="B", content="unique b")
|
||||
dups = archive.find_duplicates()
|
||||
assert dups == []
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
jan = datetime(2024, 1, 15, tzinfo=timezone.utc)
|
||||
_make_entry_at(archive, "Jan entry", jan)
|
||||
|
||||
results = archive.by_date_range("2023-01-01", "2023-12-31")
|
||||
assert results == []
|
||||
|
||||
|
||||
def test_add_skip_dups():
|
||||
def test_by_date_range_timezone_naive_treated_as_utc():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e1 = ingest_event(archive, title="Test", content="Content here")
|
||||
# Try to add exact same entry with skip_dups=True
|
||||
e2 = ArchiveEntry(title="Test", content="Content here")
|
||||
result = archive.add(e2, skip_dups=True)
|
||||
assert result.id == e1.id # returned existing, not new
|
||||
assert archive.count == 1
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
dt = datetime(2024, 6, 15, tzinfo=timezone.utc)
|
||||
e = _make_entry_at(archive, "Summer", dt)
|
||||
|
||||
# Timezone-naive start/end should still match
|
||||
results = archive.by_date_range("2024-06-01", "2024-07-01")
|
||||
assert any(r.id == e.id for r in results)
|
||||
|
||||
|
||||
def test_add_skip_dups_allows_different():
|
||||
def test_by_date_range_sorted_ascending():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "test_archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path)
|
||||
e1 = ingest_event(archive, title="A", content="Content A")
|
||||
e2 = ArchiveEntry(title="B", content="Content B")
|
||||
result = archive.add(e2, skip_dups=True)
|
||||
assert result.id == e2.id # new entry added
|
||||
assert archive.count == 2
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
dates = [
|
||||
datetime(2024, 3, 5, tzinfo=timezone.utc),
|
||||
datetime(2024, 1, 10, tzinfo=timezone.utc),
|
||||
datetime(2024, 2, 20, tzinfo=timezone.utc),
|
||||
]
|
||||
for i, dt in enumerate(dates):
|
||||
_make_entry_at(archive, f"Entry {i}", dt)
|
||||
|
||||
results = archive.by_date_range("2024-01-01", "2024-12-31")
|
||||
assert len(results) == 3
|
||||
assert results[0].created_at < results[1].created_at < results[2].created_at
|
||||
|
||||
|
||||
def test_entry_roundtrip_with_updated_at():
|
||||
e = ArchiveEntry(title="T", content="c", topics=["x"])
|
||||
d = e.to_dict()
|
||||
e2 = ArchiveEntry.from_dict(d)
|
||||
assert e2.updated_at == e.updated_at
|
||||
assert "content_hash" in d
|
||||
def test_by_date_range_single_entry_archive():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
dt = datetime(2024, 5, 1, tzinfo=timezone.utc)
|
||||
e = _make_entry_at(archive, "Only", dt)
|
||||
|
||||
assert archive.by_date_range("2024-01-01", "2024-12-31") == [e]
|
||||
assert archive.by_date_range("2025-01-01", "2025-12-31") == []
|
||||
|
||||
|
||||
def test_entry_from_dict_backfills_updated_at():
|
||||
"""Legacy entries without updated_at should get it from created_at."""
|
||||
data = {
|
||||
"id": "test-id",
|
||||
"title": "Legacy",
|
||||
"content": "old entry",
|
||||
"source": "manual",
|
||||
"source_ref": None,
|
||||
"topics": [],
|
||||
"metadata": {},
|
||||
"created_at": "2025-01-01T00:00:00+00:00",
|
||||
"links": [],
|
||||
}
|
||||
e = ArchiveEntry.from_dict(data)
|
||||
assert e.updated_at == "2025-01-01T00:00:00+00:00"
|
||||
# --- temporal_neighbors tests ---
|
||||
|
||||
def test_temporal_neighbors_empty_archive():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
e = ingest_event(archive, title="Lone", content="c")
|
||||
results = archive.temporal_neighbors(e.id, window_days=7)
|
||||
assert results == []
|
||||
|
||||
|
||||
def test_temporal_neighbors_missing_entry_raises():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
try:
|
||||
archive.temporal_neighbors("nonexistent-id")
|
||||
assert False, "Expected KeyError"
|
||||
except KeyError:
|
||||
pass
|
||||
|
||||
|
||||
def test_temporal_neighbors_returns_within_window():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
anchor_dt = datetime(2024, 4, 10, tzinfo=timezone.utc)
|
||||
near_dt = datetime(2024, 4, 14, tzinfo=timezone.utc) # +4 days — within 7
|
||||
far_dt = datetime(2024, 4, 20, tzinfo=timezone.utc) # +10 days — outside 7
|
||||
|
||||
anchor = _make_entry_at(archive, "Anchor", anchor_dt)
|
||||
near = _make_entry_at(archive, "Near", near_dt)
|
||||
far = _make_entry_at(archive, "Far", far_dt)
|
||||
|
||||
results = archive.temporal_neighbors(anchor.id, window_days=7)
|
||||
ids = {e.id for e in results}
|
||||
assert near.id in ids
|
||||
assert far.id not in ids
|
||||
assert anchor.id not in ids
|
||||
|
||||
|
||||
def test_temporal_neighbors_excludes_anchor():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
dt = datetime(2024, 4, 10, tzinfo=timezone.utc)
|
||||
anchor = _make_entry_at(archive, "Anchor", dt)
|
||||
same = _make_entry_at(archive, "Same day", dt)
|
||||
|
||||
results = archive.temporal_neighbors(anchor.id, window_days=0)
|
||||
ids = {e.id for e in results}
|
||||
assert anchor.id not in ids
|
||||
assert same.id in ids
|
||||
|
||||
|
||||
def test_temporal_neighbors_custom_window():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
anchor_dt = datetime(2024, 4, 10, tzinfo=timezone.utc)
|
||||
within_3 = datetime(2024, 4, 12, tzinfo=timezone.utc) # +2 days
|
||||
outside_3 = datetime(2024, 4, 15, tzinfo=timezone.utc) # +5 days
|
||||
|
||||
anchor = _make_entry_at(archive, "Anchor", anchor_dt)
|
||||
e_near = _make_entry_at(archive, "Near", within_3)
|
||||
e_far = _make_entry_at(archive, "Far", outside_3)
|
||||
|
||||
results = archive.temporal_neighbors(anchor.id, window_days=3)
|
||||
ids = {e.id for e in results}
|
||||
assert e_near.id in ids
|
||||
assert e_far.id not in ids
|
||||
|
||||
|
||||
def test_temporal_neighbors_sorted_ascending():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
anchor_dt = datetime(2024, 6, 15, tzinfo=timezone.utc)
|
||||
anchor = _make_entry_at(archive, "Anchor", anchor_dt)
|
||||
for offset in [5, 1, 3]:
|
||||
_make_entry_at(archive, f"Offset {offset}", anchor_dt + timedelta(days=offset))
|
||||
|
||||
results = archive.temporal_neighbors(anchor.id, window_days=7)
|
||||
assert len(results) == 3
|
||||
assert results[0].created_at < results[1].created_at < results[2].created_at
|
||||
|
||||
|
||||
def test_temporal_neighbors_boundary_inclusive():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = MnemosyneArchive(archive_path=Path(tmp) / "a.json")
|
||||
anchor_dt = datetime(2024, 6, 15, tzinfo=timezone.utc)
|
||||
boundary_dt = anchor_dt + timedelta(days=7) # exactly at window edge
|
||||
|
||||
anchor = _make_entry_at(archive, "Anchor", anchor_dt)
|
||||
boundary = _make_entry_at(archive, "Boundary", boundary_dt)
|
||||
|
||||
results = archive.temporal_neighbors(anchor.id, window_days=7)
|
||||
assert any(r.id == boundary.id for r in results)
|
||||
|
||||
138
nexus/mnemosyne/tests/test_cli_commands.py
Normal file
138
nexus/mnemosyne/tests/test_cli_commands.py
Normal file
@@ -0,0 +1,138 @@
|
||||
"""Tests for Mnemosyne CLI commands — path, touch, decay, vitality, fading, vibrant."""
|
||||
|
||||
import json
|
||||
import tempfile
|
||||
from pathlib import Path
|
||||
from unittest.mock import patch
|
||||
import sys
|
||||
import io
|
||||
|
||||
import pytest
|
||||
|
||||
from nexus.mnemosyne.archive import MnemosyneArchive
|
||||
from nexus.mnemosyne.entry import ArchiveEntry
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def archive(tmp_path):
|
||||
path = tmp_path / "test_archive.json"
|
||||
return MnemosyneArchive(archive_path=path)
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def linked_archive(tmp_path):
|
||||
"""Archive with entries linked to each other for path testing."""
|
||||
path = tmp_path / "test_archive.json"
|
||||
arch = MnemosyneArchive(archive_path=path, auto_embed=False)
|
||||
e1 = arch.add(ArchiveEntry(title="Alpha", content="first entry about python", topics=["code"]))
|
||||
e2 = arch.add(ArchiveEntry(title="Beta", content="second entry about python coding", topics=["code"]))
|
||||
e3 = arch.add(ArchiveEntry(title="Gamma", content="third entry about cooking recipes", topics=["food"]))
|
||||
return arch, e1, e2, e3
|
||||
|
||||
|
||||
class TestPathCommand:
|
||||
def test_shortest_path_exists(self, linked_archive):
|
||||
arch, e1, e2, e3 = linked_archive
|
||||
path = arch.shortest_path(e1.id, e2.id)
|
||||
assert path is not None
|
||||
assert path[0] == e1.id
|
||||
assert path[-1] == e2.id
|
||||
|
||||
def test_shortest_path_no_connection(self, linked_archive):
|
||||
arch, e1, e2, e3 = linked_archive
|
||||
# e3 (cooking) likely not linked to e1 (python coding)
|
||||
path = arch.shortest_path(e1.id, e3.id)
|
||||
# Path may or may not exist depending on linking threshold
|
||||
# Either None or a list is valid
|
||||
|
||||
def test_shortest_path_same_entry(self, linked_archive):
|
||||
arch, e1, _, _ = linked_archive
|
||||
path = arch.shortest_path(e1.id, e1.id)
|
||||
assert path == [e1.id]
|
||||
|
||||
def test_shortest_path_missing_entry(self, linked_archive):
|
||||
arch, e1, _, _ = linked_archive
|
||||
path = arch.shortest_path(e1.id, "nonexistent-id")
|
||||
assert path is None
|
||||
|
||||
|
||||
class TestTouchCommand:
|
||||
def test_touch_boosts_vitality(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content"))
|
||||
# Simulate time passing by setting old last_accessed
|
||||
old_time = "2020-01-01T00:00:00+00:00"
|
||||
entry.last_accessed = old_time
|
||||
entry.vitality = 0.5
|
||||
archive._save()
|
||||
|
||||
touched = archive.touch(entry.id)
|
||||
assert touched.vitality > 0.5
|
||||
assert touched.last_accessed != old_time
|
||||
|
||||
def test_touch_missing_entry(self, archive):
|
||||
with pytest.raises(KeyError):
|
||||
archive.touch("nonexistent-id")
|
||||
|
||||
|
||||
class TestDecayCommand:
|
||||
def test_apply_decay_returns_stats(self, archive):
|
||||
archive.add(ArchiveEntry(title="Test", content="Content"))
|
||||
result = archive.apply_decay()
|
||||
assert result["total_entries"] == 1
|
||||
assert "avg_vitality" in result
|
||||
assert "fading_count" in result
|
||||
assert "vibrant_count" in result
|
||||
|
||||
def test_decay_on_empty_archive(self, archive):
|
||||
result = archive.apply_decay()
|
||||
assert result["total_entries"] == 0
|
||||
assert result["avg_vitality"] == 0.0
|
||||
|
||||
|
||||
class TestVitalityCommand:
|
||||
def test_get_vitality(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content"))
|
||||
v = archive.get_vitality(entry.id)
|
||||
assert v["entry_id"] == entry.id
|
||||
assert v["title"] == "Test"
|
||||
assert 0.0 <= v["vitality"] <= 1.0
|
||||
assert v["age_days"] >= 0
|
||||
|
||||
def test_get_vitality_missing(self, archive):
|
||||
with pytest.raises(KeyError):
|
||||
archive.get_vitality("nonexistent-id")
|
||||
|
||||
|
||||
class TestFadingVibrant:
|
||||
def test_fading_returns_sorted_ascending(self, archive):
|
||||
# Add entries with different vitalities
|
||||
e1 = archive.add(ArchiveEntry(title="Vibrant", content="High energy"))
|
||||
e2 = archive.add(ArchiveEntry(title="Fading", content="Low energy"))
|
||||
e2.vitality = 0.1
|
||||
e2.last_accessed = "2020-01-01T00:00:00+00:00"
|
||||
archive._save()
|
||||
|
||||
results = archive.fading(limit=10)
|
||||
assert len(results) == 2
|
||||
assert results[0]["vitality"] <= results[1]["vitality"]
|
||||
|
||||
def test_vibrant_returns_sorted_descending(self, archive):
|
||||
e1 = archive.add(ArchiveEntry(title="Fresh", content="New"))
|
||||
e2 = archive.add(ArchiveEntry(title="Old", content="Ancient"))
|
||||
e2.vitality = 0.1
|
||||
e2.last_accessed = "2020-01-01T00:00:00+00:00"
|
||||
archive._save()
|
||||
|
||||
results = archive.vibrant(limit=10)
|
||||
assert len(results) == 2
|
||||
assert results[0]["vitality"] >= results[1]["vitality"]
|
||||
|
||||
def test_fading_limit(self, archive):
|
||||
for i in range(15):
|
||||
archive.add(ArchiveEntry(title=f"Entry {i}", content=f"Content {i}"))
|
||||
results = archive.fading(limit=5)
|
||||
assert len(results) == 5
|
||||
|
||||
def test_vibrant_empty(self, archive):
|
||||
results = archive.vibrant()
|
||||
assert results == []
|
||||
176
nexus/mnemosyne/tests/test_consolidation.py
Normal file
176
nexus/mnemosyne/tests/test_consolidation.py
Normal file
@@ -0,0 +1,176 @@
|
||||
"""Tests for MnemosyneArchive.consolidate() — duplicate/near-duplicate merging."""
|
||||
|
||||
import tempfile
|
||||
from pathlib import Path
|
||||
|
||||
from nexus.mnemosyne.archive import MnemosyneArchive
|
||||
from nexus.mnemosyne.entry import ArchiveEntry
|
||||
from nexus.mnemosyne.ingest import ingest_event
|
||||
|
||||
|
||||
def _archive(tmp: str) -> MnemosyneArchive:
|
||||
return MnemosyneArchive(archive_path=Path(tmp) / "archive.json", auto_embed=False)
|
||||
|
||||
|
||||
def test_consolidate_exact_duplicate_removed():
|
||||
"""Two entries with identical content_hash are merged; only one survives."""
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _archive(tmp)
|
||||
e1 = ingest_event(archive, title="Hello world", content="Exactly the same content", topics=["a"])
|
||||
# Manually add a second entry with the same hash to simulate a duplicate
|
||||
e2 = ArchiveEntry(title="Hello world", content="Exactly the same content", topics=["b"])
|
||||
# Bypass dedup guard so we can test consolidate() rather than add()
|
||||
archive._entries[e2.id] = e2
|
||||
archive._save()
|
||||
|
||||
assert archive.count == 2
|
||||
merges = archive.consolidate(dry_run=False)
|
||||
assert len(merges) == 1
|
||||
assert merges[0]["reason"] == "exact_hash"
|
||||
assert merges[0]["score"] == 1.0
|
||||
assert archive.count == 1
|
||||
|
||||
|
||||
def test_consolidate_keeps_older_entry():
|
||||
"""The older entry (earlier created_at) is kept, the newer is removed."""
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _archive(tmp)
|
||||
e1 = ingest_event(archive, title="Hello world", content="Same content here", topics=[])
|
||||
e2 = ArchiveEntry(title="Hello world", content="Same content here", topics=[])
|
||||
# Make e2 clearly newer
|
||||
e2.created_at = "2099-01-01T00:00:00+00:00"
|
||||
archive._entries[e2.id] = e2
|
||||
archive._save()
|
||||
|
||||
merges = archive.consolidate(dry_run=False)
|
||||
assert len(merges) == 1
|
||||
assert merges[0]["kept"] == e1.id
|
||||
assert merges[0]["removed"] == e2.id
|
||||
|
||||
|
||||
def test_consolidate_merges_topics():
|
||||
"""Topics from the removed entry are merged (unioned) into the kept entry."""
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _archive(tmp)
|
||||
e1 = ingest_event(archive, title="Memory item", content="Shared content body", topics=["alpha"])
|
||||
e2 = ArchiveEntry(title="Memory item", content="Shared content body", topics=["beta", "gamma"])
|
||||
e2.created_at = "2099-01-01T00:00:00+00:00"
|
||||
archive._entries[e2.id] = e2
|
||||
archive._save()
|
||||
|
||||
archive.consolidate(dry_run=False)
|
||||
survivor = archive.get(e1.id)
|
||||
assert survivor is not None
|
||||
topic_lower = {t.lower() for t in survivor.topics}
|
||||
assert "alpha" in topic_lower
|
||||
assert "beta" in topic_lower
|
||||
assert "gamma" in topic_lower
|
||||
|
||||
|
||||
def test_consolidate_merges_metadata():
|
||||
"""Metadata from the removed entry is merged into the kept entry; kept values win."""
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _archive(tmp)
|
||||
e1 = ArchiveEntry(
|
||||
title="Shared", content="Identical body here", topics=[], metadata={"k1": "v1", "shared": "kept"}
|
||||
)
|
||||
archive._entries[e1.id] = e1
|
||||
e2 = ArchiveEntry(
|
||||
title="Shared", content="Identical body here", topics=[], metadata={"k2": "v2", "shared": "removed"}
|
||||
)
|
||||
e2.created_at = "2099-01-01T00:00:00+00:00"
|
||||
archive._entries[e2.id] = e2
|
||||
archive._save()
|
||||
|
||||
archive.consolidate(dry_run=False)
|
||||
survivor = archive.get(e1.id)
|
||||
assert survivor.metadata["k1"] == "v1"
|
||||
assert survivor.metadata["k2"] == "v2"
|
||||
assert survivor.metadata["shared"] == "kept" # kept entry wins
|
||||
|
||||
|
||||
def test_consolidate_dry_run_no_mutation():
|
||||
"""Dry-run mode returns merge plan but does not alter the archive."""
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _archive(tmp)
|
||||
ingest_event(archive, title="Same", content="Identical content to dedup", topics=[])
|
||||
e2 = ArchiveEntry(title="Same", content="Identical content to dedup", topics=[])
|
||||
e2.created_at = "2099-01-01T00:00:00+00:00"
|
||||
archive._entries[e2.id] = e2
|
||||
archive._save()
|
||||
|
||||
merges = archive.consolidate(dry_run=True)
|
||||
assert len(merges) == 1
|
||||
assert merges[0]["dry_run"] is True
|
||||
# Archive must be unchanged
|
||||
assert archive.count == 2
|
||||
|
||||
|
||||
def test_consolidate_no_duplicates():
|
||||
"""When no duplicates exist, consolidate returns an empty list."""
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _archive(tmp)
|
||||
ingest_event(archive, title="Unique A", content="This is completely unique content for A")
|
||||
ingest_event(archive, title="Unique B", content="Totally different words here for B")
|
||||
merges = archive.consolidate(threshold=0.9)
|
||||
assert merges == []
|
||||
|
||||
|
||||
def test_consolidate_transfers_links():
|
||||
"""Links from the removed entry are inherited by the kept entry."""
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _archive(tmp)
|
||||
# Create a third entry to act as a link target
|
||||
target = ingest_event(archive, title="Target", content="The link target entry", topics=[])
|
||||
|
||||
e1 = ArchiveEntry(title="Dup", content="Exact duplicate body text", topics=[], links=[target.id])
|
||||
archive._entries[e1.id] = e1
|
||||
target.links.append(e1.id)
|
||||
|
||||
e2 = ArchiveEntry(title="Dup", content="Exact duplicate body text", topics=[])
|
||||
e2.created_at = "2099-01-01T00:00:00+00:00"
|
||||
archive._entries[e2.id] = e2
|
||||
archive._save()
|
||||
|
||||
archive.consolidate(dry_run=False)
|
||||
survivor = archive.get(e1.id)
|
||||
assert survivor is not None
|
||||
assert target.id in survivor.links
|
||||
|
||||
|
||||
def test_consolidate_near_duplicate_semantic():
|
||||
"""Near-duplicate entries above the similarity threshold are merged."""
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _archive(tmp)
|
||||
# Entries with very high Jaccard overlap
|
||||
text_a = "python automation scripting building tools workflows"
|
||||
text_b = "python automation scripting building tools workflows tasks"
|
||||
e1 = ArchiveEntry(title="Automator", content=text_a, topics=[])
|
||||
e2 = ArchiveEntry(title="Automator", content=text_b, topics=[])
|
||||
e2.created_at = "2099-01-01T00:00:00+00:00"
|
||||
archive._entries[e1.id] = e1
|
||||
archive._entries[e2.id] = e2
|
||||
archive._save()
|
||||
|
||||
# Use a low threshold to ensure these very similar entries match
|
||||
merges = archive.consolidate(threshold=0.7, dry_run=False)
|
||||
assert len(merges) >= 1
|
||||
assert merges[0]["reason"] == "semantic_similarity"
|
||||
|
||||
|
||||
def test_consolidate_persists_after_reload():
|
||||
"""After consolidation, the reduced archive survives a save/reload cycle."""
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "archive.json"
|
||||
archive = MnemosyneArchive(archive_path=path, auto_embed=False)
|
||||
ingest_event(archive, title="Persist test", content="Body to dedup and persist", topics=[])
|
||||
e2 = ArchiveEntry(title="Persist test", content="Body to dedup and persist", topics=[])
|
||||
e2.created_at = "2099-01-01T00:00:00+00:00"
|
||||
archive._entries[e2.id] = e2
|
||||
archive._save()
|
||||
|
||||
archive.consolidate(dry_run=False)
|
||||
assert archive.count == 1
|
||||
|
||||
reloaded = MnemosyneArchive(archive_path=path, auto_embed=False)
|
||||
assert reloaded.count == 1
|
||||
1
nexus/mnemosyne/tests/test_discover.py
Normal file
1
nexus/mnemosyne/tests/test_discover.py
Normal file
@@ -0,0 +1 @@
|
||||
# Test discover
|
||||
112
nexus/mnemosyne/tests/test_embeddings.py
Normal file
112
nexus/mnemosyne/tests/test_embeddings.py
Normal file
@@ -0,0 +1,112 @@
|
||||
"""Tests for the embedding backend module."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import math
|
||||
import pytest
|
||||
|
||||
from nexus.mnemosyne.embeddings import (
|
||||
EmbeddingBackend,
|
||||
TfidfEmbeddingBackend,
|
||||
cosine_similarity,
|
||||
get_embedding_backend,
|
||||
)
|
||||
|
||||
|
||||
class TestCosineSimilarity:
|
||||
def test_identical_vectors(self):
|
||||
a = [1.0, 2.0, 3.0]
|
||||
assert abs(cosine_similarity(a, a) - 1.0) < 1e-9
|
||||
|
||||
def test_orthogonal_vectors(self):
|
||||
a = [1.0, 0.0]
|
||||
b = [0.0, 1.0]
|
||||
assert abs(cosine_similarity(a, b) - 0.0) < 1e-9
|
||||
|
||||
def test_opposite_vectors(self):
|
||||
a = [1.0, 0.0]
|
||||
b = [-1.0, 0.0]
|
||||
assert abs(cosine_similarity(a, b) - (-1.0)) < 1e-9
|
||||
|
||||
def test_zero_vector(self):
|
||||
a = [0.0, 0.0]
|
||||
b = [1.0, 2.0]
|
||||
assert cosine_similarity(a, b) == 0.0
|
||||
|
||||
def test_dimension_mismatch(self):
|
||||
with pytest.raises(ValueError):
|
||||
cosine_similarity([1.0], [1.0, 2.0])
|
||||
|
||||
|
||||
class TestTfidfEmbeddingBackend:
|
||||
def test_basic_embed(self):
|
||||
backend = TfidfEmbeddingBackend()
|
||||
vec = backend.embed("hello world test")
|
||||
assert len(vec) > 0
|
||||
assert all(isinstance(v, float) for v in vec)
|
||||
|
||||
def test_empty_text(self):
|
||||
backend = TfidfEmbeddingBackend()
|
||||
vec = backend.embed("")
|
||||
assert vec == []
|
||||
|
||||
def test_identical_texts_similar(self):
|
||||
backend = TfidfEmbeddingBackend()
|
||||
v1 = backend.embed("the cat sat on the mat")
|
||||
v2 = backend.embed("the cat sat on the mat")
|
||||
sim = backend.similarity(v1, v2)
|
||||
assert sim > 0.99
|
||||
|
||||
def test_different_texts_less_similar(self):
|
||||
backend = TfidfEmbeddingBackend()
|
||||
v1 = backend.embed("python programming language")
|
||||
v2 = backend.embed("cooking recipes italian food")
|
||||
sim = backend.similarity(v1, v2)
|
||||
assert sim < 0.5
|
||||
|
||||
def test_related_texts_more_similar(self):
|
||||
backend = TfidfEmbeddingBackend()
|
||||
v1 = backend.embed("machine learning neural networks")
|
||||
v2 = backend.embed("deep learning artificial neural nets")
|
||||
v3 = backend.embed("baking bread sourdough recipe")
|
||||
sim_related = backend.similarity(v1, v2)
|
||||
sim_unrelated = backend.similarity(v1, v3)
|
||||
assert sim_related > sim_unrelated
|
||||
|
||||
def test_name(self):
|
||||
backend = TfidfEmbeddingBackend()
|
||||
assert "TF-IDF" in backend.name
|
||||
|
||||
def test_dimension_grows(self):
|
||||
backend = TfidfEmbeddingBackend()
|
||||
d1 = backend.dimension
|
||||
backend.embed("new unique tokens here")
|
||||
d2 = backend.dimension
|
||||
assert d2 > d1
|
||||
|
||||
def test_padding_different_lengths(self):
|
||||
backend = TfidfEmbeddingBackend()
|
||||
v1 = backend.embed("short")
|
||||
v2 = backend.embed("this is a much longer text with many more tokens")
|
||||
# Should not raise despite different lengths
|
||||
sim = backend.similarity(v1, v2)
|
||||
assert 0.0 <= sim <= 1.0
|
||||
|
||||
|
||||
class TestGetEmbeddingBackend:
|
||||
def test_tfidf_preferred(self):
|
||||
backend = get_embedding_backend(prefer="tfidf")
|
||||
assert isinstance(backend, TfidfEmbeddingBackend)
|
||||
|
||||
def test_auto_returns_something(self):
|
||||
backend = get_embedding_backend()
|
||||
assert isinstance(backend, EmbeddingBackend)
|
||||
|
||||
def test_ollama_unavailable_falls_back(self):
|
||||
# Should fall back to TF-IDF when Ollama is unreachable
|
||||
backend = get_embedding_backend(prefer="ollama", ollama_url="http://localhost:1")
|
||||
# If it raises, the test fails — it should fall back
|
||||
# But with prefer="ollama" it raises if unavailable
|
||||
# So we test without prefer:
|
||||
backend = get_embedding_backend(ollama_url="http://localhost:1")
|
||||
assert isinstance(backend, TfidfEmbeddingBackend)
|
||||
241
nexus/mnemosyne/tests/test_ingest_file.py
Normal file
241
nexus/mnemosyne/tests/test_ingest_file.py
Normal file
@@ -0,0 +1,241 @@
|
||||
"""Tests for file-based ingestion pipeline (ingest_file / ingest_directory)."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import tempfile
|
||||
from pathlib import Path
|
||||
|
||||
import pytest
|
||||
|
||||
from nexus.mnemosyne.archive import MnemosyneArchive
|
||||
from nexus.mnemosyne.ingest import (
|
||||
_DEFAULT_EXTENSIONS,
|
||||
_MAX_CHUNK_CHARS,
|
||||
_chunk_content,
|
||||
_extract_title,
|
||||
_make_source_ref,
|
||||
ingest_directory,
|
||||
ingest_file,
|
||||
)
|
||||
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# Helpers
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
def _make_archive(tmp_path: Path) -> MnemosyneArchive:
|
||||
return MnemosyneArchive(archive_path=tmp_path / "archive.json")
|
||||
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# Unit: _extract_title
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
def test_extract_title_from_heading():
|
||||
content = "# My Document\n\nSome content here."
|
||||
assert _extract_title(content, Path("ignored.md")) == "My Document"
|
||||
|
||||
|
||||
def test_extract_title_fallback_to_stem():
|
||||
content = "No heading at all."
|
||||
assert _extract_title(content, Path("/docs/my_notes.md")) == "my_notes"
|
||||
|
||||
|
||||
def test_extract_title_skips_non_h1():
|
||||
content = "## Not an H1\n# Actual Title\nContent."
|
||||
assert _extract_title(content, Path("x.md")) == "Actual Title"
|
||||
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# Unit: _make_source_ref
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
def test_source_ref_format():
|
||||
p = Path("/tmp/foo.md")
|
||||
ref = _make_source_ref(p, 1234567890.9)
|
||||
assert ref == "file:/tmp/foo.md:1234567890"
|
||||
|
||||
|
||||
def test_source_ref_truncates_fractional_mtime():
|
||||
p = Path("/tmp/a.txt")
|
||||
assert _make_source_ref(p, 100.99) == _make_source_ref(p, 100.01)
|
||||
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# Unit: _chunk_content
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
def test_chunk_short_content_is_single():
|
||||
content = "Short content."
|
||||
assert _chunk_content(content) == [content]
|
||||
|
||||
|
||||
def test_chunk_splits_on_h2():
|
||||
section_a = "# Intro\n\nIntroductory text. " + "x" * 100
|
||||
section_b = "## Section B\n\nBody of section B. " + "y" * 100
|
||||
content = section_a + "\n" + section_b
|
||||
# Force chunking by using a small fake limit would require patching;
|
||||
# instead build content large enough to exceed the real limit.
|
||||
big_a = "# Intro\n\n" + "a" * (_MAX_CHUNK_CHARS - 50)
|
||||
big_b = "## Section B\n\n" + "b" * (_MAX_CHUNK_CHARS - 50)
|
||||
combined = big_a + "\n" + big_b
|
||||
chunks = _chunk_content(combined)
|
||||
assert len(chunks) >= 2
|
||||
assert any("Section B" in c for c in chunks)
|
||||
|
||||
|
||||
def test_chunk_fixed_window_fallback():
|
||||
# Content with no ## headings but > MAX_CHUNK_CHARS
|
||||
content = "word " * (_MAX_CHUNK_CHARS // 5 + 100)
|
||||
chunks = _chunk_content(content)
|
||||
assert len(chunks) >= 2
|
||||
for c in chunks:
|
||||
assert len(c) <= _MAX_CHUNK_CHARS
|
||||
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# ingest_file
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
def test_ingest_file_returns_entry(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
doc = tmp_path / "notes.md"
|
||||
doc.write_text("# My Notes\n\nHello world.")
|
||||
entries = ingest_file(archive, doc)
|
||||
assert len(entries) == 1
|
||||
assert entries[0].title == "My Notes"
|
||||
assert entries[0].source == "file"
|
||||
assert "Hello world" in entries[0].content
|
||||
|
||||
|
||||
def test_ingest_file_uses_stem_when_no_heading(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
doc = tmp_path / "raw_log.txt"
|
||||
doc.write_text("Just some plain text without a heading.")
|
||||
entries = ingest_file(archive, doc)
|
||||
assert entries[0].title == "raw_log"
|
||||
|
||||
|
||||
def test_ingest_file_dedup_unchanged(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
doc = tmp_path / "doc.md"
|
||||
doc.write_text("# Title\n\nContent.")
|
||||
entries1 = ingest_file(archive, doc)
|
||||
assert archive.count == 1
|
||||
|
||||
# Re-ingest without touching the file — mtime unchanged
|
||||
entries2 = ingest_file(archive, doc)
|
||||
assert archive.count == 1 # no duplicate
|
||||
assert entries2[0].id == entries1[0].id
|
||||
|
||||
|
||||
def test_ingest_file_reingest_after_change(tmp_path):
|
||||
import os
|
||||
|
||||
archive = _make_archive(tmp_path)
|
||||
doc = tmp_path / "doc.md"
|
||||
doc.write_text("# Title\n\nOriginal content.")
|
||||
ingest_file(archive, doc)
|
||||
assert archive.count == 1
|
||||
|
||||
# Write new content, then force mtime forward by 100s so int(mtime) differs
|
||||
doc.write_text("# Title\n\nUpdated content.")
|
||||
new_mtime = doc.stat().st_mtime + 100
|
||||
os.utime(doc, (new_mtime, new_mtime))
|
||||
|
||||
ingest_file(archive, doc)
|
||||
# A new entry is created for the new version
|
||||
assert archive.count == 2
|
||||
|
||||
|
||||
def test_ingest_file_source_ref_contains_path(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
doc = tmp_path / "thing.txt"
|
||||
doc.write_text("Plain text.")
|
||||
entries = ingest_file(archive, doc)
|
||||
assert str(doc) in entries[0].source_ref
|
||||
|
||||
|
||||
def test_ingest_file_large_produces_chunks(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
doc = tmp_path / "big.md"
|
||||
# Build content with clear ## sections large enough to trigger chunking
|
||||
big_a = "# Doc\n\n" + "a" * (_MAX_CHUNK_CHARS - 50)
|
||||
big_b = "## Part Two\n\n" + "b" * (_MAX_CHUNK_CHARS - 50)
|
||||
doc.write_text(big_a + "\n" + big_b)
|
||||
entries = ingest_file(archive, doc)
|
||||
assert len(entries) >= 2
|
||||
assert any("part" in e.title.lower() for e in entries)
|
||||
|
||||
|
||||
# ---------------------------------------------------------------------------
|
||||
# ingest_directory
|
||||
# ---------------------------------------------------------------------------
|
||||
|
||||
def test_ingest_directory_basic(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
docs = tmp_path / "docs"
|
||||
docs.mkdir()
|
||||
(docs / "a.md").write_text("# Alpha\n\nFirst doc.")
|
||||
(docs / "b.txt").write_text("Beta plain text.")
|
||||
(docs / "skip.py").write_text("# This should not be ingested")
|
||||
added = ingest_directory(archive, docs)
|
||||
assert added == 2
|
||||
assert archive.count == 2
|
||||
|
||||
|
||||
def test_ingest_directory_custom_extensions(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
docs = tmp_path / "docs"
|
||||
docs.mkdir()
|
||||
(docs / "a.md").write_text("# Alpha")
|
||||
(docs / "b.py").write_text("No heading — uses stem.")
|
||||
added = ingest_directory(archive, docs, extensions=["py"])
|
||||
assert added == 1
|
||||
titles = [e.title for e in archive._entries.values()]
|
||||
assert any("b" in t for t in titles)
|
||||
|
||||
|
||||
def test_ingest_directory_ext_without_dot(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
docs = tmp_path / "docs"
|
||||
docs.mkdir()
|
||||
(docs / "notes.md").write_text("# Notes\n\nContent.")
|
||||
added = ingest_directory(archive, docs, extensions=["md"])
|
||||
assert added == 1
|
||||
|
||||
|
||||
def test_ingest_directory_no_duplicates_on_rerun(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
docs = tmp_path / "docs"
|
||||
docs.mkdir()
|
||||
(docs / "file.md").write_text("# Stable\n\nSame content.")
|
||||
ingest_directory(archive, docs)
|
||||
assert archive.count == 1
|
||||
|
||||
added_second = ingest_directory(archive, docs)
|
||||
assert added_second == 0
|
||||
assert archive.count == 1
|
||||
|
||||
|
||||
def test_ingest_directory_recurses_subdirs(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
docs = tmp_path / "docs"
|
||||
sub = docs / "sub"
|
||||
sub.mkdir(parents=True)
|
||||
(docs / "top.md").write_text("# Top level")
|
||||
(sub / "nested.md").write_text("# Nested")
|
||||
added = ingest_directory(archive, docs)
|
||||
assert added == 2
|
||||
|
||||
|
||||
def test_ingest_directory_default_extensions(tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
docs = tmp_path / "docs"
|
||||
docs.mkdir()
|
||||
(docs / "a.md").write_text("markdown")
|
||||
(docs / "b.txt").write_text("text")
|
||||
(docs / "c.json").write_text('{"key": "value"}')
|
||||
(docs / "d.yaml").write_text("key: value")
|
||||
added = ingest_directory(archive, docs)
|
||||
assert added == 3 # md, txt, json — not yaml
|
||||
278
nexus/mnemosyne/tests/test_memory_decay.py
Normal file
278
nexus/mnemosyne/tests/test_memory_decay.py
Normal file
@@ -0,0 +1,278 @@
|
||||
"""Tests for Mnemosyne memory decay system."""
|
||||
|
||||
import json
|
||||
import os
|
||||
import tempfile
|
||||
from datetime import datetime, timedelta, timezone
|
||||
from pathlib import Path
|
||||
|
||||
import pytest
|
||||
|
||||
from nexus.mnemosyne.archive import MnemosyneArchive
|
||||
from nexus.mnemosyne.entry import ArchiveEntry
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def archive(tmp_path):
|
||||
"""Create a fresh archive for testing."""
|
||||
path = tmp_path / "test_archive.json"
|
||||
return MnemosyneArchive(archive_path=path)
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def populated_archive(tmp_path):
|
||||
"""Create an archive with some entries."""
|
||||
path = tmp_path / "test_archive.json"
|
||||
arch = MnemosyneArchive(archive_path=path)
|
||||
arch.add(ArchiveEntry(title="Fresh Entry", content="Just added", topics=["test"]))
|
||||
arch.add(ArchiveEntry(title="Old Entry", content="Been here a while", topics=["test"]))
|
||||
arch.add(ArchiveEntry(title="Another Entry", content="Some content", topics=["other"]))
|
||||
return arch
|
||||
|
||||
|
||||
class TestVitalityFields:
|
||||
"""Test that vitality fields exist on entries."""
|
||||
|
||||
def test_entry_has_vitality_default(self):
|
||||
entry = ArchiveEntry(title="Test", content="Content")
|
||||
assert entry.vitality == 1.0
|
||||
|
||||
def test_entry_has_last_accessed_default(self):
|
||||
entry = ArchiveEntry(title="Test", content="Content")
|
||||
assert entry.last_accessed is None
|
||||
|
||||
def test_entry_roundtrip_with_vitality(self):
|
||||
entry = ArchiveEntry(
|
||||
title="Test", content="Content",
|
||||
vitality=0.75,
|
||||
last_accessed="2024-01-01T00:00:00+00:00"
|
||||
)
|
||||
d = entry.to_dict()
|
||||
assert d["vitality"] == 0.75
|
||||
assert d["last_accessed"] == "2024-01-01T00:00:00+00:00"
|
||||
restored = ArchiveEntry.from_dict(d)
|
||||
assert restored.vitality == 0.75
|
||||
assert restored.last_accessed == "2024-01-01T00:00:00+00:00"
|
||||
|
||||
|
||||
class TestTouch:
|
||||
"""Test touch() access recording and vitality boost."""
|
||||
|
||||
def test_touch_sets_last_accessed(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content"))
|
||||
assert entry.last_accessed is None
|
||||
touched = archive.touch(entry.id)
|
||||
assert touched.last_accessed is not None
|
||||
|
||||
def test_touch_boosts_vitality(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content", vitality=0.5))
|
||||
touched = archive.touch(entry.id)
|
||||
# Boost = 0.1 * (1 - 0.5) = 0.05, so vitality should be ~0.55
|
||||
# (assuming no time decay in test — instantaneous)
|
||||
assert touched.vitality > 0.5
|
||||
assert touched.vitality <= 1.0
|
||||
|
||||
def test_touch_diminishing_returns(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content", vitality=0.9))
|
||||
touched = archive.touch(entry.id)
|
||||
# Boost = 0.1 * (1 - 0.9) = 0.01, so vitality should be ~0.91
|
||||
assert touched.vitality < 0.92
|
||||
assert touched.vitality > 0.9
|
||||
|
||||
def test_touch_never_exceeds_one(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content", vitality=0.99))
|
||||
for _ in range(10):
|
||||
entry = archive.touch(entry.id)
|
||||
assert entry.vitality <= 1.0
|
||||
|
||||
def test_touch_missing_entry_raises(self, archive):
|
||||
with pytest.raises(KeyError):
|
||||
archive.touch("nonexistent-id")
|
||||
|
||||
def test_touch_persists(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content"))
|
||||
archive.touch(entry.id)
|
||||
# Reload archive
|
||||
arch2 = MnemosyneArchive(archive_path=archive._path)
|
||||
loaded = arch2.get(entry.id)
|
||||
assert loaded.last_accessed is not None
|
||||
|
||||
|
||||
class TestGetVitality:
|
||||
"""Test get_vitality() status reporting."""
|
||||
|
||||
def test_get_vitality_basic(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content"))
|
||||
status = archive.get_vitality(entry.id)
|
||||
assert status["entry_id"] == entry.id
|
||||
assert status["title"] == "Test"
|
||||
assert 0.0 <= status["vitality"] <= 1.0
|
||||
assert status["age_days"] == 0
|
||||
|
||||
def test_get_vitality_missing_raises(self, archive):
|
||||
with pytest.raises(KeyError):
|
||||
archive.get_vitality("nonexistent-id")
|
||||
|
||||
|
||||
class TestComputeVitality:
|
||||
"""Test the decay computation."""
|
||||
|
||||
def test_new_entry_full_vitality(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content"))
|
||||
v = archive._compute_vitality(entry)
|
||||
assert v == 1.0
|
||||
|
||||
def test_recently_touched_high_vitality(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content"))
|
||||
archive.touch(entry.id)
|
||||
v = archive._compute_vitality(entry)
|
||||
assert v > 0.99 # Should be essentially 1.0 since just touched
|
||||
|
||||
def test_old_entry_decays(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content"))
|
||||
# Simulate old access — set last_accessed to 60 days ago
|
||||
old_date = (datetime.now(timezone.utc) - timedelta(days=60)).isoformat()
|
||||
entry.last_accessed = old_date
|
||||
entry.vitality = 1.0
|
||||
archive._save()
|
||||
v = archive._compute_vitality(entry)
|
||||
# 60 days with 30-day half-life: v = 1.0 * 0.5^(60/30) = 0.25
|
||||
assert v < 0.3
|
||||
assert v > 0.2
|
||||
|
||||
def test_very_old_entry_nearly_zero(self, archive):
|
||||
entry = archive.add(ArchiveEntry(title="Test", content="Content"))
|
||||
old_date = (datetime.now(timezone.utc) - timedelta(days=365)).isoformat()
|
||||
entry.last_accessed = old_date
|
||||
entry.vitality = 1.0
|
||||
archive._save()
|
||||
v = archive._compute_vitality(entry)
|
||||
# 365 days / 30 half-life = ~12 half-lives -> ~0.0002
|
||||
assert v < 0.01
|
||||
|
||||
|
||||
class TestFading:
|
||||
"""Test fading() — most neglected entries."""
|
||||
|
||||
def test_fading_returns_lowest_first(self, populated_archive):
|
||||
entries = list(populated_archive._entries.values())
|
||||
# Make one entry very old
|
||||
old_entry = entries[1]
|
||||
old_date = (datetime.now(timezone.utc) - timedelta(days=90)).isoformat()
|
||||
old_entry.last_accessed = old_date
|
||||
old_entry.vitality = 1.0
|
||||
populated_archive._save()
|
||||
|
||||
fading = populated_archive.fading(limit=3)
|
||||
assert len(fading) <= 3
|
||||
# First result should be the oldest
|
||||
assert fading[0]["entry_id"] == old_entry.id
|
||||
# Should be in ascending order
|
||||
for i in range(len(fading) - 1):
|
||||
assert fading[i]["vitality"] <= fading[i + 1]["vitality"]
|
||||
|
||||
def test_fading_empty_archive(self, archive):
|
||||
fading = archive.fading()
|
||||
assert fading == []
|
||||
|
||||
def test_fading_limit(self, populated_archive):
|
||||
fading = populated_archive.fading(limit=2)
|
||||
assert len(fading) == 2
|
||||
|
||||
|
||||
class TestVibrant:
|
||||
"""Test vibrant() — most alive entries."""
|
||||
|
||||
def test_vibrant_returns_highest_first(self, populated_archive):
|
||||
entries = list(populated_archive._entries.values())
|
||||
# Make one entry very old
|
||||
old_entry = entries[1]
|
||||
old_date = (datetime.now(timezone.utc) - timedelta(days=90)).isoformat()
|
||||
old_entry.last_accessed = old_date
|
||||
old_entry.vitality = 1.0
|
||||
populated_archive._save()
|
||||
|
||||
vibrant = populated_archive.vibrant(limit=3)
|
||||
# Should be in descending order
|
||||
for i in range(len(vibrant) - 1):
|
||||
assert vibrant[i]["vitality"] >= vibrant[i + 1]["vitality"]
|
||||
# First result should NOT be the old entry
|
||||
assert vibrant[0]["entry_id"] != old_entry.id
|
||||
|
||||
def test_vibrant_empty_archive(self, archive):
|
||||
vibrant = archive.vibrant()
|
||||
assert vibrant == []
|
||||
|
||||
|
||||
class TestApplyDecay:
|
||||
"""Test apply_decay() bulk decay operation."""
|
||||
|
||||
def test_apply_decay_returns_stats(self, populated_archive):
|
||||
result = populated_archive.apply_decay()
|
||||
assert result["total_entries"] == 3
|
||||
assert "decayed_count" in result
|
||||
assert "avg_vitality" in result
|
||||
assert "fading_count" in result
|
||||
assert "vibrant_count" in result
|
||||
|
||||
def test_apply_decay_persists(self, populated_archive):
|
||||
populated_archive.apply_decay()
|
||||
# Reload
|
||||
arch2 = MnemosyneArchive(archive_path=populated_archive._path)
|
||||
result2 = arch2.apply_decay()
|
||||
# Should show same entries
|
||||
assert result2["total_entries"] == 3
|
||||
|
||||
def test_apply_decay_on_empty(self, archive):
|
||||
result = archive.apply_decay()
|
||||
assert result["total_entries"] == 0
|
||||
assert result["avg_vitality"] == 0.0
|
||||
|
||||
|
||||
class TestStatsVitality:
|
||||
"""Test that stats() includes vitality summary."""
|
||||
|
||||
def test_stats_includes_vitality(self, populated_archive):
|
||||
stats = populated_archive.stats()
|
||||
assert "avg_vitality" in stats
|
||||
assert "fading_count" in stats
|
||||
assert "vibrant_count" in stats
|
||||
assert 0.0 <= stats["avg_vitality"] <= 1.0
|
||||
|
||||
def test_stats_empty_archive(self, archive):
|
||||
stats = archive.stats()
|
||||
assert stats["avg_vitality"] == 0.0
|
||||
assert stats["fading_count"] == 0
|
||||
assert stats["vibrant_count"] == 0
|
||||
|
||||
|
||||
class TestDecayLifecycle:
|
||||
"""Integration test: full lifecycle from creation to fading."""
|
||||
|
||||
def test_entry_lifecycle(self, archive):
|
||||
# Create
|
||||
entry = archive.add(ArchiveEntry(title="Memory", content="A thing happened"))
|
||||
assert entry.vitality == 1.0
|
||||
|
||||
# Touch a few times
|
||||
for _ in range(5):
|
||||
archive.touch(entry.id)
|
||||
|
||||
# Check it's vibrant
|
||||
vibrant = archive.vibrant(limit=1)
|
||||
assert len(vibrant) == 1
|
||||
assert vibrant[0]["entry_id"] == entry.id
|
||||
|
||||
# Simulate time passing
|
||||
entry.last_accessed = (datetime.now(timezone.utc) - timedelta(days=45)).isoformat()
|
||||
entry.vitality = 0.8
|
||||
archive._save()
|
||||
|
||||
# Apply decay
|
||||
result = archive.apply_decay()
|
||||
assert result["total_entries"] == 1
|
||||
|
||||
# Check it's now fading
|
||||
fading = archive.fading(limit=1)
|
||||
assert fading[0]["entry_id"] == entry.id
|
||||
assert fading[0]["vitality"] < 0.5
|
||||
106
nexus/mnemosyne/tests/test_path.py
Normal file
106
nexus/mnemosyne/tests/test_path.py
Normal file
@@ -0,0 +1,106 @@
|
||||
"""Tests for MnemosyneArchive.shortest_path and path_explanation."""
|
||||
|
||||
from nexus.mnemosyne.archive import MnemosyneArchive
|
||||
from nexus.mnemosyne.entry import ArchiveEntry
|
||||
|
||||
|
||||
def _make_archive(tmp_path):
|
||||
archive = MnemosyneArchive(str(tmp_path / "test_archive.json"))
|
||||
return archive
|
||||
|
||||
|
||||
class TestShortestPath:
|
||||
def test_direct_connection(self, tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
a = archive.add("Alpha", "first entry", topics=["start"])
|
||||
b = archive.add("Beta", "second entry", topics=["end"])
|
||||
# Manually link
|
||||
a.links.append(b.id)
|
||||
b.links.append(a.id)
|
||||
archive._entries[a.id] = a
|
||||
archive._entries[b.id] = b
|
||||
archive._save()
|
||||
|
||||
path = archive.shortest_path(a.id, b.id)
|
||||
assert path == [a.id, b.id]
|
||||
|
||||
def test_multi_hop_path(self, tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
a = archive.add("A", "alpha", topics=["x"])
|
||||
b = archive.add("B", "beta", topics=["y"])
|
||||
c = archive.add("C", "gamma", topics=["z"])
|
||||
# Chain: A -> B -> C
|
||||
a.links.append(b.id)
|
||||
b.links.extend([a.id, c.id])
|
||||
c.links.append(b.id)
|
||||
archive._entries[a.id] = a
|
||||
archive._entries[b.id] = b
|
||||
archive._entries[c.id] = c
|
||||
archive._save()
|
||||
|
||||
path = archive.shortest_path(a.id, c.id)
|
||||
assert path == [a.id, b.id, c.id]
|
||||
|
||||
def test_no_path(self, tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
a = archive.add("A", "isolated", topics=[])
|
||||
b = archive.add("B", "also isolated", topics=[])
|
||||
path = archive.shortest_path(a.id, b.id)
|
||||
assert path is None
|
||||
|
||||
def test_same_entry(self, tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
a = archive.add("A", "lonely", topics=[])
|
||||
path = archive.shortest_path(a.id, a.id)
|
||||
assert path == [a.id]
|
||||
|
||||
def test_nonexistent_entry(self, tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
a = archive.add("A", "exists", topics=[])
|
||||
path = archive.shortest_path("fake-id", a.id)
|
||||
assert path is None
|
||||
|
||||
def test_shortest_of_multiple(self, tmp_path):
|
||||
"""When multiple paths exist, BFS returns shortest."""
|
||||
archive = _make_archive(tmp_path)
|
||||
a = archive.add("A", "a", topics=[])
|
||||
b = archive.add("B", "b", topics=[])
|
||||
c = archive.add("C", "c", topics=[])
|
||||
d = archive.add("D", "d", topics=[])
|
||||
# A -> B -> D (short)
|
||||
# A -> C -> B -> D (long)
|
||||
a.links.extend([b.id, c.id])
|
||||
b.links.extend([a.id, d.id, c.id])
|
||||
c.links.extend([a.id, b.id])
|
||||
d.links.append(b.id)
|
||||
for e in [a, b, c, d]:
|
||||
archive._entries[e.id] = e
|
||||
archive._save()
|
||||
|
||||
path = archive.shortest_path(a.id, d.id)
|
||||
assert len(path) == 3 # A -> B -> D, not A -> C -> B -> D
|
||||
|
||||
|
||||
class TestPathExplanation:
|
||||
def test_returns_step_details(self, tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
a = archive.add("Alpha", "the beginning", topics=["origin"])
|
||||
b = archive.add("Beta", "the middle", topics=["process"])
|
||||
a.links.append(b.id)
|
||||
b.links.append(a.id)
|
||||
archive._entries[a.id] = a
|
||||
archive._entries[b.id] = b
|
||||
archive._save()
|
||||
|
||||
path = [a.id, b.id]
|
||||
steps = archive.path_explanation(path)
|
||||
assert len(steps) == 2
|
||||
assert steps[0]["title"] == "Alpha"
|
||||
assert steps[1]["title"] == "Beta"
|
||||
assert "origin" in steps[0]["topics"]
|
||||
|
||||
def test_content_preview_truncation(self, tmp_path):
|
||||
archive = _make_archive(tmp_path)
|
||||
a = archive.add("A", "x" * 200, topics=[])
|
||||
steps = archive.path_explanation([a.id])
|
||||
assert len(steps[0]["content_preview"]) <= 123 # 120 + "..."
|
||||
1
nexus/mnemosyne/tests/test_resonance.py
Normal file
1
nexus/mnemosyne/tests/test_resonance.py
Normal file
@@ -0,0 +1 @@
|
||||
# Test resonance
|
||||
1
nexus/mnemosyne/tests/test_snapshot.py
Normal file
1
nexus/mnemosyne/tests/test_snapshot.py
Normal file
@@ -0,0 +1 @@
|
||||
# Test snapshot
|
||||
240
nexus/mnemosyne/tests/test_snapshots.py
Normal file
240
nexus/mnemosyne/tests/test_snapshots.py
Normal file
@@ -0,0 +1,240 @@
|
||||
"""Tests for Mnemosyne snapshot (point-in-time backup/restore) feature."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import json
|
||||
import tempfile
|
||||
from pathlib import Path
|
||||
|
||||
import pytest
|
||||
|
||||
from nexus.mnemosyne.archive import MnemosyneArchive
|
||||
from nexus.mnemosyne.ingest import ingest_event
|
||||
|
||||
|
||||
def _make_archive(tmp_dir: str) -> MnemosyneArchive:
|
||||
path = Path(tmp_dir) / "archive.json"
|
||||
return MnemosyneArchive(archive_path=path, auto_embed=False)
|
||||
|
||||
|
||||
# ─── snapshot_create ─────────────────────────────────────────────────────────
|
||||
|
||||
def test_snapshot_create_returns_metadata():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
ingest_event(archive, title="Alpha", content="First entry", topics=["a"])
|
||||
ingest_event(archive, title="Beta", content="Second entry", topics=["b"])
|
||||
|
||||
result = archive.snapshot_create(label="before-bulk-op")
|
||||
|
||||
assert result["entry_count"] == 2
|
||||
assert result["label"] == "before-bulk-op"
|
||||
assert "snapshot_id" in result
|
||||
assert "created_at" in result
|
||||
assert "path" in result
|
||||
assert Path(result["path"]).exists()
|
||||
|
||||
|
||||
def test_snapshot_create_no_label():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
ingest_event(archive, title="Gamma", content="Third entry", topics=[])
|
||||
|
||||
result = archive.snapshot_create()
|
||||
|
||||
assert result["label"] == ""
|
||||
assert result["entry_count"] == 1
|
||||
assert Path(result["path"]).exists()
|
||||
|
||||
|
||||
def test_snapshot_file_contains_entries():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
e = ingest_event(archive, title="Delta", content="Fourth entry", topics=["d"])
|
||||
result = archive.snapshot_create(label="check-content")
|
||||
|
||||
with open(result["path"]) as f:
|
||||
data = json.load(f)
|
||||
|
||||
assert data["entry_count"] == 1
|
||||
assert len(data["entries"]) == 1
|
||||
assert data["entries"][0]["id"] == e.id
|
||||
assert data["entries"][0]["title"] == "Delta"
|
||||
|
||||
|
||||
def test_snapshot_create_empty_archive():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
result = archive.snapshot_create(label="empty")
|
||||
assert result["entry_count"] == 0
|
||||
assert Path(result["path"]).exists()
|
||||
|
||||
|
||||
# ─── snapshot_list ───────────────────────────────────────────────────────────
|
||||
|
||||
def test_snapshot_list_empty():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
assert archive.snapshot_list() == []
|
||||
|
||||
|
||||
def test_snapshot_list_returns_all():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
ingest_event(archive, title="One", content="c1", topics=[])
|
||||
archive.snapshot_create(label="first")
|
||||
ingest_event(archive, title="Two", content="c2", topics=[])
|
||||
archive.snapshot_create(label="second")
|
||||
|
||||
snapshots = archive.snapshot_list()
|
||||
assert len(snapshots) == 2
|
||||
labels = {s["label"] for s in snapshots}
|
||||
assert "first" in labels
|
||||
assert "second" in labels
|
||||
|
||||
|
||||
def test_snapshot_list_metadata_fields():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
archive.snapshot_create(label="meta-check")
|
||||
snapshots = archive.snapshot_list()
|
||||
s = snapshots[0]
|
||||
for key in ("snapshot_id", "label", "created_at", "entry_count", "path"):
|
||||
assert key in s
|
||||
|
||||
|
||||
def test_snapshot_list_newest_first():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
archive.snapshot_create(label="a")
|
||||
archive.snapshot_create(label="b")
|
||||
snapshots = archive.snapshot_list()
|
||||
# Filenames sort lexicographically; newest (b) should be first
|
||||
# (filenames include timestamp so alphabetical = newest-last;
|
||||
# snapshot_list reverses the glob order → newest first)
|
||||
assert len(snapshots) == 2
|
||||
# Both should be present; ordering is newest first
|
||||
ids = [s["snapshot_id"] for s in snapshots]
|
||||
assert ids == sorted(ids, reverse=True)
|
||||
|
||||
|
||||
# ─── snapshot_restore ────────────────────────────────────────────────────────
|
||||
|
||||
def test_snapshot_restore_replaces_entries():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
ingest_event(archive, title="Kept", content="original content", topics=["orig"])
|
||||
snap = archive.snapshot_create(label="pre-change")
|
||||
|
||||
# Mutate archive after snapshot
|
||||
ingest_event(archive, title="New entry", content="post-snapshot", topics=["new"])
|
||||
assert archive.count == 2
|
||||
|
||||
result = archive.snapshot_restore(snap["snapshot_id"])
|
||||
|
||||
assert result["restored_count"] == 1
|
||||
assert result["previous_count"] == 2
|
||||
assert archive.count == 1
|
||||
entry = list(archive._entries.values())[0]
|
||||
assert entry.title == "Kept"
|
||||
|
||||
|
||||
def test_snapshot_restore_persists_to_disk():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
path = Path(tmp) / "archive.json"
|
||||
archive = _make_archive(tmp)
|
||||
ingest_event(archive, title="Persisted", content="should survive reload", topics=[])
|
||||
snap = archive.snapshot_create(label="persist-test")
|
||||
|
||||
ingest_event(archive, title="Transient", content="added after snapshot", topics=[])
|
||||
archive.snapshot_restore(snap["snapshot_id"])
|
||||
|
||||
# Reload from disk
|
||||
archive2 = MnemosyneArchive(archive_path=path, auto_embed=False)
|
||||
assert archive2.count == 1
|
||||
assert list(archive2._entries.values())[0].title == "Persisted"
|
||||
|
||||
|
||||
def test_snapshot_restore_missing_raises():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
with pytest.raises(FileNotFoundError):
|
||||
archive.snapshot_restore("nonexistent_snapshot_id")
|
||||
|
||||
|
||||
# ─── snapshot_diff ───────────────────────────────────────────────────────────
|
||||
|
||||
def test_snapshot_diff_no_changes():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
ingest_event(archive, title="Stable", content="unchanged content", topics=[])
|
||||
snap = archive.snapshot_create(label="baseline")
|
||||
|
||||
diff = archive.snapshot_diff(snap["snapshot_id"])
|
||||
|
||||
assert diff["added"] == []
|
||||
assert diff["removed"] == []
|
||||
assert diff["modified"] == []
|
||||
assert diff["unchanged"] == 1
|
||||
|
||||
|
||||
def test_snapshot_diff_detects_added():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
ingest_event(archive, title="Original", content="existing", topics=[])
|
||||
snap = archive.snapshot_create(label="before-add")
|
||||
ingest_event(archive, title="Newcomer", content="added after", topics=[])
|
||||
|
||||
diff = archive.snapshot_diff(snap["snapshot_id"])
|
||||
|
||||
assert len(diff["added"]) == 1
|
||||
assert diff["added"][0]["title"] == "Newcomer"
|
||||
assert diff["removed"] == []
|
||||
assert diff["unchanged"] == 1
|
||||
|
||||
|
||||
def test_snapshot_diff_detects_removed():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
e1 = ingest_event(archive, title="Will Be Removed", content="doomed", topics=[])
|
||||
ingest_event(archive, title="Survivor", content="stays", topics=[])
|
||||
snap = archive.snapshot_create(label="pre-removal")
|
||||
archive.remove(e1.id)
|
||||
|
||||
diff = archive.snapshot_diff(snap["snapshot_id"])
|
||||
|
||||
assert len(diff["removed"]) == 1
|
||||
assert diff["removed"][0]["title"] == "Will Be Removed"
|
||||
assert diff["added"] == []
|
||||
assert diff["unchanged"] == 1
|
||||
|
||||
|
||||
def test_snapshot_diff_detects_modified():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
e = ingest_event(archive, title="Mutable", content="original content", topics=[])
|
||||
snap = archive.snapshot_create(label="pre-edit")
|
||||
archive.update_entry(e.id, content="updated content", auto_link=False)
|
||||
|
||||
diff = archive.snapshot_diff(snap["snapshot_id"])
|
||||
|
||||
assert len(diff["modified"]) == 1
|
||||
assert diff["modified"][0]["title"] == "Mutable"
|
||||
assert diff["modified"][0]["snapshot_hash"] != diff["modified"][0]["current_hash"]
|
||||
assert diff["added"] == []
|
||||
assert diff["removed"] == []
|
||||
|
||||
|
||||
def test_snapshot_diff_missing_raises():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
with pytest.raises(FileNotFoundError):
|
||||
archive.snapshot_diff("no_such_snapshot")
|
||||
|
||||
|
||||
def test_snapshot_diff_includes_snapshot_id():
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
archive = _make_archive(tmp)
|
||||
snap = archive.snapshot_create(label="id-check")
|
||||
diff = archive.snapshot_diff(snap["snapshot_id"])
|
||||
assert diff["snapshot_id"] == snap["snapshot_id"]
|
||||
@@ -1,27 +1,5 @@
|
||||
#!/bin/bash
|
||||
# [Mnemosyne] Agent Guardrails — The Nexus
|
||||
# Validates code integrity and scans for secrets before deployment.
|
||||
|
||||
echo "--- [Mnemosyne] Running Guardrails ---"
|
||||
|
||||
# 1. Syntax Checks
|
||||
echo "[1/3] Validating syntax..."
|
||||
for f in ; do
|
||||
node --check "$f" || { echo "Syntax error in $f"; exit 1; }
|
||||
done
|
||||
echo "Syntax OK."
|
||||
|
||||
# 2. JSON/YAML Validation
|
||||
echo "[2/3] Validating configs..."
|
||||
for f in ; do
|
||||
node -e "JSON.parse(require('fs').readFileSync('$f'))" || { echo "Invalid JSON: $f"; exit 1; }
|
||||
done
|
||||
echo "Configs OK."
|
||||
|
||||
# 3. Secret Scan
|
||||
echo "[3/3] Scanning for secrets..."
|
||||
grep -rE "AI_|TOKEN|KEY|SECRET" . --exclude-dir=node_modules --exclude=guardrails.sh | grep -v "process.env" && {
|
||||
echo "WARNING: Potential secrets found!"
|
||||
} || echo "No secrets detected."
|
||||
|
||||
echo "--- Guardrails Passed ---"
|
||||
echo "Running GOFAI guardrails..."
|
||||
# Syntax checks
|
||||
find . -name "*.js" -exec node --check {} +
|
||||
echo "Guardrails passed."
|
||||
|
||||
@@ -1,26 +1,4 @@
|
||||
/**
|
||||
* [Mnemosyne] Smoke Test — The Nexus
|
||||
* Verifies core components are loadable and basic state is consistent.
|
||||
*/
|
||||
|
||||
import { SpatialMemory } from '../nexus/components/spatial-memory.js';
|
||||
import { MemoryOptimizer } from '../nexus/components/memory-optimizer.js';
|
||||
|
||||
console.log('--- [Mnemosyne] Running Smoke Test ---');
|
||||
|
||||
// 1. Verify Components
|
||||
if (!SpatialMemory || !MemoryOptimizer) {
|
||||
console.error('Failed to load core components');
|
||||
process.exit(1);
|
||||
}
|
||||
console.log('Components loaded.');
|
||||
|
||||
// 2. Verify Regions
|
||||
const regions = Object.keys(SpatialMemory.REGIONS || {});
|
||||
if (regions.length < 5) {
|
||||
console.error('SpatialMemory regions incomplete:', regions);
|
||||
process.exit(1);
|
||||
}
|
||||
console.log('Regions verified:', regions.join(', '));
|
||||
|
||||
console.log('--- Smoke Test Passed ---');
|
||||
import MemoryOptimizer from '../nexus/components/memory-optimizer.js';
|
||||
const optimizer = new MemoryOptimizer();
|
||||
console.log('Smoke test passed');
|
||||
|
||||
160
style.css
160
style.css
@@ -1917,3 +1917,163 @@ canvas#nexus-canvas {
|
||||
background: rgba(74, 240, 192, 0.18);
|
||||
border-color: #4af0c0;
|
||||
}
|
||||
|
||||
/* ═══ MNEMOSYNE: Memory Connections Panel ═══ */
|
||||
.memory-connections-panel {
|
||||
position: fixed;
|
||||
top: 50%;
|
||||
right: 280px;
|
||||
transform: translateY(-50%) translateX(12px);
|
||||
width: 260px;
|
||||
max-height: 70vh;
|
||||
background: rgba(10, 12, 18, 0.92);
|
||||
border: 1px solid rgba(74, 240, 192, 0.15);
|
||||
border-radius: 8px;
|
||||
box-shadow: 0 8px 32px rgba(0,0,0,0.5);
|
||||
z-index: 310;
|
||||
display: flex;
|
||||
flex-direction: column;
|
||||
opacity: 0;
|
||||
transition: opacity 0.2s ease, transform 0.2s ease;
|
||||
backdrop-filter: blur(8px);
|
||||
-webkit-backdrop-filter: blur(8px);
|
||||
font-family: var(--font-mono, monospace);
|
||||
}
|
||||
.memory-connections-panel.mc-visible {
|
||||
opacity: 1;
|
||||
transform: translateY(-50%) translateX(0);
|
||||
}
|
||||
|
||||
.mc-header {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: space-between;
|
||||
padding: 10px 14px;
|
||||
border-bottom: 1px solid rgba(255, 255, 255, 0.06);
|
||||
}
|
||||
.mc-title {
|
||||
color: rgba(74, 240, 192, 0.8);
|
||||
font-size: 11px;
|
||||
font-weight: 600;
|
||||
text-transform: uppercase;
|
||||
letter-spacing: 0.5px;
|
||||
}
|
||||
.mc-close {
|
||||
background: none;
|
||||
border: none;
|
||||
color: rgba(255, 255, 255, 0.4);
|
||||
font-size: 14px;
|
||||
cursor: pointer;
|
||||
padding: 2px 6px;
|
||||
border-radius: 4px;
|
||||
line-height: 1;
|
||||
}
|
||||
.mc-close:hover {
|
||||
color: #fff;
|
||||
background: rgba(255, 255, 255, 0.1);
|
||||
}
|
||||
|
||||
.mc-section {
|
||||
padding: 8px 14px 10px;
|
||||
border-bottom: 1px solid rgba(255, 255, 255, 0.05);
|
||||
}
|
||||
.mc-section:last-child { border-bottom: none; }
|
||||
|
||||
.mc-section-label {
|
||||
color: rgba(74, 240, 192, 0.5);
|
||||
font-size: 9px;
|
||||
text-transform: uppercase;
|
||||
letter-spacing: 1px;
|
||||
margin-bottom: 8px;
|
||||
}
|
||||
|
||||
.mc-conn-list, .mc-suggest-list {
|
||||
max-height: 200px;
|
||||
overflow-y: auto;
|
||||
}
|
||||
.mc-conn-list::-webkit-scrollbar, .mc-suggest-list::-webkit-scrollbar { width: 3px; }
|
||||
.mc-conn-list::-webkit-scrollbar-thumb, .mc-suggest-list::-webkit-scrollbar-thumb {
|
||||
background: rgba(74, 240, 192, 0.15);
|
||||
border-radius: 2px;
|
||||
}
|
||||
|
||||
.mc-conn-item, .mc-suggest-item {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: space-between;
|
||||
padding: 6px 8px;
|
||||
border-radius: 5px;
|
||||
margin-bottom: 4px;
|
||||
transition: background 0.15s ease;
|
||||
}
|
||||
.mc-conn-item:hover {
|
||||
background: rgba(74, 240, 192, 0.06);
|
||||
}
|
||||
.mc-suggest-item:hover {
|
||||
background: rgba(123, 92, 255, 0.06);
|
||||
}
|
||||
|
||||
.mc-conn-info, .mc-suggest-info {
|
||||
flex: 1;
|
||||
min-width: 0;
|
||||
overflow: hidden;
|
||||
}
|
||||
.mc-conn-label, .mc-suggest-label {
|
||||
display: block;
|
||||
color: var(--color-text, #ccc);
|
||||
font-size: 11px;
|
||||
white-space: nowrap;
|
||||
overflow: hidden;
|
||||
text-overflow: ellipsis;
|
||||
}
|
||||
.mc-conn-meta, .mc-suggest-meta {
|
||||
display: block;
|
||||
color: rgba(255, 255, 255, 0.3);
|
||||
font-size: 9px;
|
||||
margin-top: 1px;
|
||||
}
|
||||
|
||||
.mc-conn-actions {
|
||||
display: flex;
|
||||
gap: 4px;
|
||||
flex-shrink: 0;
|
||||
margin-left: 8px;
|
||||
}
|
||||
|
||||
.mc-btn {
|
||||
background: none;
|
||||
border: 1px solid rgba(255, 255, 255, 0.12);
|
||||
color: rgba(255, 255, 255, 0.5);
|
||||
cursor: pointer;
|
||||
border-radius: 4px;
|
||||
font-size: 12px;
|
||||
padding: 2px 6px;
|
||||
line-height: 1;
|
||||
transition: all 0.15s ease;
|
||||
}
|
||||
.mc-btn-nav:hover {
|
||||
border-color: #4af0c0;
|
||||
color: #4af0c0;
|
||||
background: rgba(74, 240, 192, 0.08);
|
||||
}
|
||||
.mc-btn-remove:hover {
|
||||
border-color: #ff4466;
|
||||
color: #ff4466;
|
||||
background: rgba(255, 68, 102, 0.08);
|
||||
}
|
||||
.mc-btn-add {
|
||||
border-color: rgba(123, 92, 255, 0.3);
|
||||
color: rgba(123, 92, 255, 0.7);
|
||||
}
|
||||
.mc-btn-add:hover {
|
||||
border-color: #7b5cff;
|
||||
color: #7b5cff;
|
||||
background: rgba(123, 92, 255, 0.12);
|
||||
}
|
||||
|
||||
.mc-empty {
|
||||
color: rgba(255, 255, 255, 0.25);
|
||||
font-size: 11px;
|
||||
font-style: italic;
|
||||
padding: 4px 0;
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user