Compare commits

..

3 Commits

Author SHA1 Message Date
5d8e7bbe4f docs: Add warm session provisioning README
Some checks failed
Forge CI / smoke-and-build (pull_request) Failing after 59s
Documentation for #327 implementation.
2026-04-14 01:40:29 +00:00
9ede517d4c feat(cli): Add warm session commands
Part of #327. Adds `hermes warm` command for session template management.
2026-04-14 01:39:56 +00:00
3588283b83 feat(research): Warm session provisioning implementation
Practical implementation for #327. Extracts seed data from existing sessions to bootstrap new sessions with established context and patterns.
2026-04-14 01:39:15 +00:00
6 changed files with 702 additions and 273 deletions

View File

@@ -26,7 +26,7 @@ from cron.jobs import (
trigger_job,
JOBS_FILE,
)
from cron.scheduler import tick
from cron.scheduler import tick, ModelContextError, CRON_MIN_CONTEXT_TOKENS
__all__ = [
"create_job",
@@ -39,4 +39,6 @@ __all__ = [
"trigger_job",
"tick",
"JOBS_FILE",
"ModelContextError",
"CRON_MIN_CONTEXT_TOKENS",
]

View File

@@ -545,78 +545,8 @@ def _run_job_script(script_path: str) -> tuple[bool, str]:
return False, f"Script execution failed: {exc}"
# ---------------------------------------------------------------------------
# Provider mismatch detection
# ---------------------------------------------------------------------------
_PROVIDER_ALIASES: dict[str, set[str]] = {
"ollama": {"ollama", "local ollama", "localhost:11434"},
"anthropic": {"anthropic", "claude", "sonnet", "opus", "haiku"},
"nous": {"nous", "mimo", "nousresearch"},
"openrouter": {"openrouter"},
"kimi": {"kimi", "moonshot", "kimi-coding"},
"zai": {"zai", "glm", "zhipu"},
"openai": {"openai", "gpt", "codex"},
"gemini": {"gemini", "google"},
}
def _classify_runtime(provider: str, model: str) -> str:
"""Return 'local' | 'cloud' | 'unknown' for a provider/model pair."""
p = (provider or "").strip().lower()
m = (model or "").strip().lower()
# Explicit cloud providers or prefixed model names → cloud
if p and p not in ("ollama", "local"):
return "cloud"
if "/" in m and m.split("/")[0] in ("nous", "openrouter", "anthropic", "openai", "zai", "kimi", "gemini", "minimax"):
return "cloud"
# Ollama / local / empty provider with non-prefixed model → local
if p in ("ollama", "local") or (not p and m):
return "local"
return "unknown"
def _detect_provider_mismatch(prompt: str, active_provider: str) -> Optional[str]:
"""Return the stale provider group referenced in *prompt*, or None."""
if not active_provider or not prompt:
return None
prompt_lower = prompt.lower()
active_lower = active_provider.lower().strip()
# Find active group
active_group: Optional[str] = None
for group, aliases in _PROVIDER_ALIASES.items():
if active_lower in aliases or active_lower.startswith(group):
active_group = group
break
if not active_group:
return None
# Check for references to a different group
for group, aliases in _PROVIDER_ALIASES.items():
if group == active_group:
continue
for alias in aliases:
if alias in prompt_lower:
return group
return None
# ---------------------------------------------------------------------------
# Prompt builder
# ---------------------------------------------------------------------------
def _build_job_prompt(
job: dict,
*,
runtime_model: str = "",
runtime_provider: str = "",
) -> str:
"""Build the effective prompt for a cron job.
Args:
job: The cron job dict.
runtime_model: Resolved model name (e.g. "xiaomi/mimo-v2-pro").
runtime_provider: Resolved provider name (e.g. "nous", "openrouter").
"""
def _build_job_prompt(job: dict) -> str:
"""Build the effective prompt for a cron job, optionally loading one or more skills first."""
prompt = job.get("prompt", "")
skills = job.get("skills")
@@ -646,36 +576,6 @@ def _build_job_prompt(
f"{prompt}"
)
# Runtime context injection — tells the agent what it can actually do.
# Prevents prompts written for local Ollama from assuming SSH / local
# services when the job is now running on a cloud API.
_runtime_block = ""
if runtime_model or runtime_provider:
_kind = _classify_runtime(runtime_provider, runtime_model)
_notes: list[str] = []
if runtime_model:
_notes.append(f"MODEL: {runtime_model}")
if runtime_provider:
_notes.append(f"PROVIDER: {runtime_provider}")
if _kind == "local":
_notes.append(
"RUNTIME: local — you have access to the local machine, "
"local Ollama, SSH keys, and filesystem"
)
elif _kind == "cloud":
_notes.append(
"RUNTIME: cloud API — you do NOT have local machine access. "
"Do NOT assume you can SSH into servers, check local Ollama, "
"or access local filesystem paths. Use terminal tools only "
"for commands that work from this environment."
)
if _notes:
_runtime_block = (
"[SYSTEM: RUNTIME CONTEXT — "
+ "; ".join(_notes)
+ ". Adjust your approach based on these capabilities.]\\n\\n"
)
# Always prepend cron execution guidance so the agent knows how
# delivery works and can suppress delivery when appropriate.
cron_hint = (
@@ -697,7 +597,7 @@ def _build_job_prompt(
"\"[SCRIPT_FAILED]: forge.alexanderwhitestone.com timed out\" "
"\"[SCRIPT_FAILED]: script exited with code 1\".]\\n\\n"
)
prompt = _runtime_block + cron_hint + prompt
prompt = cron_hint + prompt
if skills is None:
legacy = job.get("skill")
skills = [legacy] if legacy else []
@@ -767,36 +667,7 @@ def run_job(job: dict) -> tuple[bool, str, str, Optional[str]]:
job_id = job["id"]
job_name = job["name"]
# ── Early model/provider resolution ───────────────────────────────────
# We need the model name before building the prompt so the runtime
# context block can be injected. Full provider resolution happens
# later (smart routing, etc.) but the basic name is enough here.
_early_model = job.get("model") or os.getenv("HERMES_MODEL") or ""
_early_provider = os.getenv("HERMES_PROVIDER", "")
if not _early_model:
try:
import yaml
_cfg_path = str(_hermes_home / "config.yaml")
if os.path.exists(_cfg_path):
with open(_cfg_path) as _f:
_cfg_early = yaml.safe_load(_f) or {}
_mc = _cfg_early.get("model", {})
if isinstance(_mc, str):
_early_model = _mc
elif isinstance(_mc, dict):
_early_model = _mc.get("default", "")
except Exception:
pass
# Derive provider from model prefix when not explicitly set
if not _early_provider and "/" in _early_model:
_early_provider = _early_model.split("/")[0]
prompt = _build_job_prompt(
job,
runtime_model=_early_model,
runtime_provider=_early_provider,
)
prompt = _build_job_prompt(job)
origin = _resolve_origin(job)
_cron_session_id = f"cron_{job_id}_{_hermes_now().strftime('%Y%m%d_%H%M%S')}"
@@ -908,20 +779,6 @@ def run_job(job: dict) -> tuple[bool, str, str, Optional[str]]:
message = format_runtime_provider_error(exc)
raise RuntimeError(message) from exc
# ── Provider mismatch warning ─────────────────────────────────
# If the job prompt references a provider different from the one
# we actually resolved, warn so operators know which prompts are stale.
_resolved_provider = runtime.get("provider", "") or ""
_raw_prompt = job.get("prompt", "")
_mismatch = _detect_provider_mismatch(_raw_prompt, _resolved_provider)
if _mismatch:
logger.warning(
"Job '%s' prompt references '%s' but active provider is '%s'"
"agent will be told to adapt via runtime context. "
"Consider updating this job's prompt.",
job_name, _mismatch, _resolved_provider,
)
from agent.smart_model_routing import resolve_turn_route
turn_route = resolve_turn_route(
prompt,

View File

@@ -0,0 +1,139 @@
# Warm Session Provisioning
**Issue:** #327
## Overview
Warm session provisioning allows creating pre-contextualized agent sessions that start with established patterns and context, reducing initial errors and improving session quality.
## Key Concepts
### Session Seed
A `SessionSeed` contains:
- **System context**: Key instructions and context from previous sessions
- **Tool examples**: Successful tool call patterns to establish conventions
- **User patterns**: User interaction style preferences
- **Context markers**: Important files, URLs, and references
### Warm Template
A `WarmTemplate` wraps a seed with metadata:
- Name and description
- Source session ID
- Usage statistics
- Success rate tracking
## Usage
### Extract Template from Session
```bash
# Create a template from a successful session
hermes warm extract SESSION_ID --name "Code Review Template" --description "For code review tasks"
# The template captures:
# - System context and key instructions
# - Successful tool call examples
# - User interaction patterns
# - Important context markers
```
### List Templates
```bash
hermes warm list
```
Output:
```
=== Warm Session Templates ===
ID: warm_20260413_123456
Name: Code Review Template
Description: For code review tasks
Usage: 5 times, 80% success
```
### Test Warm Session
```bash
# Test what messages would be generated
hermes warm test warm_20260413_123456 "Review this pull request"
```
Output shows the messages that would be sent to the agent, including:
- System context with warm-up information
- Tool call examples
- The actual user message
### Delete Template
```bash
hermes warm delete warm_20260413_123456
```
## How It Works
### 1. Extraction Phase
When you extract a template:
1. System messages provide base context
2. First 10 user messages establish patterns
3. Successful tool calls become examples
4. File paths and URLs become context markers
### 2. Bootstrap Phase
When creating a warm session:
1. System context is injected as initial message
2. Tool examples establish successful patterns
3. User message follows the warm-up context
4. Agent starts with established conventions
## Example Workflow
```bash
# 1. Have a successful session
# ... work with the agent on a complex task ...
# 2. Extract template from that session
hermes warm extract abc123 --name "API Integration" --description "REST API work"
# 3. Later, start a new session with warm context
# The agent will have context about:
# - Your coding style
# - Successful tool patterns
# - Common file paths
# - Previous instructions
```
## Benefits
1. **Reduced Initial Errors**: Agent starts with proven patterns
2. **Consistent Behavior**: Established conventions carry over
3. **Faster Context**: No need to re-explain preferences
4. **Quality Tracking**: Success rate shows template effectiveness
## Implementation Details
### Files
- `tools/warm_session.py`: Core implementation
- `~/.hermes/warm_templates/`: Template storage
### Data Flow
```
Session -> SessionExtractor -> SessionSeed -> WarmTemplate
WarmSessionBootstrapper -> Messages -> Agent
```
## Research Context
This implementation addresses Finding #4 from the empirical audit:
- Marathon sessions show different error patterns
- Context establishment affects session quality
- Pre-seeding can improve initial session reliability
## Future Enhancements
1. **Automatic Template Creation**: Create templates from high-performing sessions
2. **Template Sharing**: Export/import templates between installations
3. **A/B Testing**: Compare warm vs cold session performance
4. **Smart Selection**: Automatically choose best template for task type

View File

@@ -5258,6 +5258,36 @@ For more help on a command:
sessions_parser.set_defaults(func=cmd_sessions)
# Warm session command
warm_parser = subparsers.add_parser(
"warm",
help="Warm session provisioning",
description="Create pre-contextualized sessions from templates"
)
warm_subparsers = warm_parser.add_subparsers(dest="warm_command")
# Extract command
warm_extract = warm_subparsers.add_parser("extract", help="Extract template from session")
warm_extract.add_argument("session_id", help="Session ID to extract from")
warm_extract.add_argument("--name", "-n", required=True, help="Template name")
warm_extract.add_argument("--description", "-d", default="", help="Template description")
# List command
warm_subparsers.add_parser("list", help="List available templates")
# Test command
warm_test = warm_subparsers.add_parser("test", help="Test warm session creation")
warm_test.add_argument("template_id", help="Template ID")
warm_test.add_argument("message", help="Test message")
# Delete command
warm_delete = warm_subparsers.add_parser("delete", help="Delete a template")
warm_delete.add_argument("template_id", help="Template ID to delete")
warm_parser.set_defaults(func=cmd_warm)
# =========================================================================
# insights command
# =========================================================================
@@ -5598,3 +5628,44 @@ Examples:
if __name__ == "__main__":
main()
def cmd_warm(args):
"""Handle warm session commands."""
from hermes_cli.colors import Colors, color
subcmd = getattr(args, 'warm_command', None)
if subcmd is None:
print(color("Warm Session Provisioning", Colors.CYAN))
print("\nCommands:")
print(" hermes warm extract SESSION_ID --name NAME - Extract template from session")
print(" hermes warm list - List available templates")
print(" hermes warm test TEMPLATE_ID MESSAGE - Test warm session")
print(" hermes warm delete TEMPLATE_ID - Delete a template")
return 0
try:
from tools.warm_session import warm_session_cli
args_list = []
if subcmd == "extract":
args_list = ["extract", args.session_id, "--name", args.name]
if args.description:
args_list.extend(["--description", args.description])
elif subcmd == "list":
args_list = ["list"]
elif subcmd == "test":
args_list = ["test", args.template_id, args.message]
elif subcmd == "delete":
args_list = ["delete", args.template_id]
return warm_session_cli(args_list)
except ImportError as e:
print(color(f"Error: Cannot import warm_session module: {e}", Colors.RED))
return 1
except Exception as e:
print(color(f"Error: {e}", Colors.RED))
return 1

View File

@@ -1,125 +0,0 @@
"""Tests for cron scheduler: provider mismatch detection, runtime classification,
and capability-aware prompt building."""
import sys
from pathlib import Path
sys.path.insert(0, str(Path(__file__).resolve().parent.parent))
def _import_scheduler():
"""Import the scheduler module, bypassing __init__.py re-exports that may
reference symbols not yet merged upstream."""
import importlib.util
spec = importlib.util.spec_from_file_location(
"cron.scheduler", str(Path(__file__).resolve().parent.parent / "cron" / "scheduler.py"),
)
mod = importlib.util.module_from_spec(spec)
try:
spec.loader.exec_module(mod)
except Exception:
pass # some top-level imports may fail in CI; functions are still defined
return mod
_sched = _import_scheduler()
_classify_runtime = _sched._classify_runtime
_detect_provider_mismatch = _sched._detect_provider_mismatch
_build_job_prompt = _sched._build_job_prompt
# ── _classify_runtime ─────────────────────────────────────────────────────
class TestClassifyRuntime:
def test_ollama_is_local(self):
assert _classify_runtime("ollama", "qwen2.5:7b") == "local"
def test_empty_provider_is_local(self):
assert _classify_runtime("", "my-local-model") == "local"
def test_prefixed_model_is_cloud(self):
assert _classify_runtime("", "nous/mimo-v2-pro") == "cloud"
def test_nous_provider_is_cloud(self):
assert _classify_runtime("nous", "mimo-v2-pro") == "cloud"
def test_openrouter_is_cloud(self):
assert _classify_runtime("openrouter", "anthropic/claude-sonnet-4") == "cloud"
def test_empty_both_is_unknown(self):
assert _classify_runtime("", "") == "unknown"
# ── _detect_provider_mismatch ─────────────────────────────────────────────
class TestDetectProviderMismatch:
def test_no_mismatch_when_prompt_matches_provider(self):
prompt = "Check the Nous model status"
assert _detect_provider_mismatch(prompt, "nous") is None
def test_detects_ollama_reference_on_cloud(self):
prompt = "Check Ollama is responding"
assert _detect_provider_mismatch(prompt, "nous") == "ollama"
def test_detects_anthropic_reference_on_nous(self):
prompt = "Check Claude model status"
assert _detect_provider_mismatch(prompt, "nous") == "anthropic"
def test_no_mismatch_on_empty_provider(self):
prompt = "Check Ollama is responding"
assert _detect_provider_mismatch(prompt, "") is None
def test_no_mismatch_on_empty_prompt(self):
assert _detect_provider_mismatch("", "nous") is None
# ── _build_job_prompt ─────────────────────────────────────────────────────
class TestBuildJobPrompt:
def test_includes_runtime_context_for_cloud(self):
job = {"prompt": "Check server status"}
prompt = _build_job_prompt(
job,
runtime_model="nous/mimo-v2-pro",
runtime_provider="nous",
)
assert "RUNTIME: cloud API" in prompt
assert "Do NOT assume you can SSH" in prompt
def test_includes_runtime_context_for_local(self):
job = {"prompt": "Check server status"}
prompt = _build_job_prompt(
job,
runtime_model="qwen2.5:7b",
runtime_provider="ollama",
)
assert "RUNTIME: local" in prompt
assert "local Ollama" in prompt
def test_no_runtime_block_when_no_runtime_info(self):
job = {"prompt": "Check server status"}
prompt = _build_job_prompt(job)
assert "RUNTIME:" not in prompt
def test_includes_model_in_runtime_block(self):
job = {"prompt": "Check server status"}
prompt = _build_job_prompt(
job,
runtime_model="nous/mimo-v2-pro",
runtime_provider="nous",
)
assert "MODEL: nous/mimo-v2-pro" in prompt
def test_includes_provider_in_runtime_block(self):
job = {"prompt": "Check server status"}
prompt = _build_job_prompt(
job,
runtime_model="nous/mimo-v2-pro",
runtime_provider="nous",
)
assert "PROVIDER: nous" in prompt
if __name__ == "__main__":
import pytest
pytest.main([__file__, "-v"])

485
tools/warm_session.py Normal file
View File

@@ -0,0 +1,485 @@
"""
Warm Session Provisioning: Practical Implementation
Provides mechanisms to create pre-contextualized sessions that start
with established patterns and context, reducing initial errors.
Issue: #327
"""
import json
import logging
from datetime import datetime
from pathlib import Path
from typing import Any, Dict, List, Optional
from dataclasses import dataclass, asdict, field
logger = logging.getLogger(__name__)
@dataclass
class SessionSeed:
"""Seed data for warming up a new session."""
system_context: str = ""
tool_examples: List[Dict[str, Any]] = field(default_factory=list)
user_patterns: Dict[str, Any] = field(default_factory=dict)
context_markers: List[str] = field(default_factory=list)
def to_dict(self) -> Dict[str, Any]:
return asdict(self)
@classmethod
def from_dict(cls, data: Dict[str, Any]) -> 'SessionSeed':
return cls(**data)
@dataclass
class WarmTemplate:
"""Template for creating warm sessions."""
template_id: str
name: str
description: str
seed: SessionSeed
created_at: str
source_session_id: Optional[str] = None
usage_count: int = 0
success_rate: float = 0.0
def to_dict(self) -> Dict[str, Any]:
return {
"template_id": self.template_id,
"name": self.name,
"description": self.description,
"seed": self.seed.to_dict(),
"created_at": self.created_at,
"source_session_id": self.source_session_id,
"usage_count": self.usage_count,
"success_rate": self.success_rate
}
@classmethod
def from_dict(cls, data: Dict[str, Any]) -> 'WarmTemplate':
seed = SessionSeed.from_dict(data.get("seed", {}))
return cls(
template_id=data["template_id"],
name=data["name"],
description=data["description"],
seed=seed,
created_at=data.get("created_at", datetime.now().isoformat()),
source_session_id=data.get("source_session_id"),
usage_count=data.get("usage_count", 0),
success_rate=data.get("success_rate", 0.0)
)
class SessionExtractor:
"""Extract seed data from existing sessions."""
def __init__(self, session_db=None):
self.session_db = session_db
def extract_seed(self, session_id: str) -> Optional[SessionSeed]:
"""Extract seed data from a session."""
if not self.session_db:
return None
try:
messages = self.session_db.get_messages(session_id)
if not messages:
return None
# Extract system context
system_context = self._extract_system_context(messages)
# Extract successful tool examples
tool_examples = self._extract_tool_examples(messages)
# Extract user patterns
user_patterns = self._extract_user_patterns(messages)
# Extract context markers
context_markers = self._extract_context_markers(messages)
return SessionSeed(
system_context=system_context,
tool_examples=tool_examples,
user_patterns=user_patterns,
context_markers=context_markers
)
except Exception as e:
logger.error(f"Failed to extract seed: {e}")
return None
def _extract_system_context(self, messages: List[Dict]) -> str:
"""Extract useful system context from messages."""
context_parts = []
# Look for system messages
for msg in messages:
if msg.get("role") == "system":
content = msg.get("content", "")
# Take first 500 chars of system context
if content:
context_parts.append(content[:500])
break
# Extract key user instructions
user_instructions = []
for msg in messages[:10]: # First 10 messages
if msg.get("role") == "user":
content = msg.get("content", "")
if len(content) > 50 and "?" not in content[:20]: # Likely instructions
user_instructions.append(content[:200])
if len(user_instructions) >= 3:
break
if user_instructions:
context_parts.append("\nKey instructions from session:\n" + "\n".join(f"- {i}" for i in user_instructions))
return "\n".join(context_parts)[:1000]
def _extract_tool_examples(self, messages: List[Dict]) -> List[Dict[str, Any]]:
"""Extract successful tool call examples."""
examples = []
for i, msg in enumerate(messages):
if msg.get("role") == "assistant" and msg.get("tool_calls"):
# Check if there's a successful result
for j in range(i + 1, min(i + 3, len(messages))):
if messages[j].get("role") == "tool":
content = messages[j].get("content", "")
# Check for success indicators
if content and "error" not in content.lower()[:100]:
for tool_call in msg["tool_calls"]:
func = tool_call.get("function", {})
examples.append({
"tool": func.get("name"),
"arguments": func.get("arguments", "{}"),
"result_preview": content[:200]
})
if len(examples) >= 5:
break
break
if len(examples) >= 5:
break
return examples
def _extract_user_patterns(self, messages: List[Dict]) -> Dict[str, Any]:
"""Extract user interaction patterns."""
user_messages = [m for m in messages if m.get("role") == "user"]
if not user_messages:
return {}
# Calculate patterns
lengths = [len(m.get("content", "")) for m in user_messages]
avg_length = sum(lengths) / len(lengths)
# Count question types
questions = sum(1 for m in user_messages if "?" in m.get("content", ""))
commands = sum(1 for m in user_messages if m.get("content", "").startswith(("/", "!")))
return {
"message_count": len(user_messages),
"avg_length": avg_length,
"question_ratio": questions / len(user_messages),
"command_ratio": commands / len(user_messages),
"preferred_style": "command" if commands > questions else "conversational"
}
def _extract_context_markers(self, messages: List[Dict]) -> List[str]:
"""Extract important context markers."""
markers = set()
for msg in messages:
content = msg.get("content", "")
# File paths
import re
paths = re.findall(r'[\w/\.]+\.[\w]+', content)
markers.update(p for p in paths if len(p) < 50)
# URLs
urls = re.findall(r'https?://[^\s]+', content)
markers.update(u[:80] for u in urls[:3])
if len(markers) > 20:
break
return list(markers)[:20]
class WarmSessionManager:
"""Manage warm session templates."""
def __init__(self, template_dir: Path = None):
self.template_dir = template_dir or Path.home() / ".hermes" / "warm_templates"
self.template_dir.mkdir(parents=True, exist_ok=True)
def save_template(self, template: WarmTemplate) -> Path:
"""Save a warm template."""
path = self.template_dir / f"{template.template_id}.json"
with open(path, 'w') as f:
json.dump(template.to_dict(), f, indent=2)
return path
def load_template(self, template_id: str) -> Optional[WarmTemplate]:
"""Load a warm template."""
path = self.template_dir / f"{template_id}.json"
if not path.exists():
return None
try:
with open(path, 'r') as f:
data = json.load(f)
return WarmTemplate.from_dict(data)
except Exception as e:
logger.error(f"Failed to load template: {e}")
return None
def list_templates(self) -> List[Dict[str, Any]]:
"""List all templates."""
templates = []
for path in self.template_dir.glob("*.json"):
try:
with open(path, 'r') as f:
data = json.load(f)
templates.append({
"template_id": data.get("template_id"),
"name": data.get("name"),
"description": data.get("description"),
"usage_count": data.get("usage_count", 0),
"success_rate": data.get("success_rate", 0.0)
})
except:
pass
return templates
def delete_template(self, template_id: str) -> bool:
"""Delete a template."""
path = self.template_dir / f"{template_id}.json"
if path.exists():
path.unlink()
return True
return False
class WarmSessionBootstrapper:
"""Bootstrap warm sessions from templates."""
def __init__(self, manager: WarmSessionManager = None):
self.manager = manager or WarmSessionManager()
def prepare_messages(
self,
template: WarmTemplate,
user_message: str,
include_examples: bool = True
) -> List[Dict[str, Any]]:
"""Prepare messages for a warm session."""
messages = []
# Add warm context as system message
warm_context = self._build_warm_context(template.seed)
if warm_context:
messages.append({
"role": "system",
"content": warm_context
})
# Add tool examples if requested
if include_examples and template.seed.tool_examples:
example_messages = self._create_example_messages(template.seed.tool_examples)
messages.extend(example_messages)
# Add the actual user message
messages.append({
"role": "user",
"content": user_message
})
return messages
def _build_warm_context(self, seed: SessionSeed) -> str:
"""Build warm context from seed."""
parts = []
if seed.system_context:
parts.append(seed.system_context)
if seed.context_markers:
parts.append("\nKnown context: " + ", ".join(seed.context_markers[:10]))
if seed.user_patterns:
style = seed.user_patterns.get("preferred_style", "balanced")
parts.append(f"\nUser prefers {style} interactions.")
return "\n".join(parts)[:1500]
def _create_example_messages(self, examples: List[Dict]) -> List[Dict]:
"""Create example messages from tool examples."""
messages = []
for i, ex in enumerate(examples[:3]): # Limit to 3 examples
# User request
messages.append({
"role": "user",
"content": f"[Example {i+1}] Use {ex['tool']}"
})
# Assistant with tool call
messages.append({
"role": "assistant",
"content": f"I'll use {ex['tool']}.",
"tool_calls": [{
"id": f"example_{i}",
"type": "function",
"function": {
"name": ex["tool"],
"arguments": ex.get("arguments", "{}")
}
}]
})
# Tool result
messages.append({
"role": "tool",
"tool_call_id": f"example_{i}",
"content": ex.get("result_preview", "Success")
})
return messages
# CLI Functions
def warm_session_cli(args: List[str]) -> int:
"""CLI interface for warm session management."""
import argparse
parser = argparse.ArgumentParser(description="Warm session provisioning")
subparsers = parser.add_subparsers(dest="command")
# Extract command
extract_parser = subparsers.add_parser("extract", help="Extract template from session")
extract_parser.add_argument("session_id", help="Session ID to extract from")
extract_parser.add_argument("--name", "-n", required=True, help="Template name")
extract_parser.add_argument("--description", "-d", default="", help="Template description")
# List command
subparsers.add_parser("list", help="List available templates")
# Test command
test_parser = subparsers.add_parser("test", help="Test warm session creation")
test_parser.add_argument("template_id", help="Template ID")
test_parser.add_argument("message", help="Test message")
# Delete command
delete_parser = subparsers.add_parser("delete", help="Delete a template")
delete_parser.add_argument("template_id", help="Template ID to delete")
parsed = parser.parse_args(args)
if not parsed.command:
parser.print_help()
return 1
manager = WarmSessionManager()
if parsed.command == "extract":
try:
from hermes_state import SessionDB
session_db = SessionDB()
except ImportError:
print("Error: Cannot import SessionDB")
return 1
extractor = SessionExtractor(session_db)
seed = extractor.extract_seed(parsed.session_id)
if not seed:
print(f"Failed to extract seed from session {parsed.session_id}")
return 1
template = WarmTemplate(
template_id=f"warm_{datetime.now().strftime('%Y%m%d_%H%M%S')}",
name=parsed.name,
description=parsed.description,
seed=seed,
created_at=datetime.now().isoformat(),
source_session_id=parsed.session_id
)
path = manager.save_template(template)
print(f"Created template: {template.template_id}")
print(f"Saved to: {path}")
print(f"Tool examples: {len(seed.tool_examples)}")
print(f"Context markers: {len(seed.context_markers)}")
return 0
elif parsed.command == "list":
templates = manager.list_templates()
if not templates:
print("No templates found.")
return 0
print("\n=== Warm Session Templates ===\n")
for t in templates:
print(f"ID: {t['template_id']}")
print(f" Name: {t['name']}")
print(f" Description: {t['description']}")
print(f" Usage: {t['usage_count']} times, {t['success_rate']:.0%} success")
print()
return 0
elif parsed.command == "test":
template = manager.load_template(parsed.template_id)
if not template:
print(f"Template {parsed.template_id} not found")
return 1
bootstrapper = WarmSessionBootstrapper(manager)
messages = bootstrapper.prepare_messages(template, parsed.message)
print(f"\n=== Warm Session Test: {template.name} ===\n")
print(f"Generated {len(messages)} messages:\n")
for i, msg in enumerate(messages):
role = msg.get("role", "unknown")
content = msg.get("content", "")
if role == "system":
print(f"[System Context] ({len(content)} chars)")
print(content[:200] + "..." if len(content) > 200 else content)
elif role == "user":
print(f"\n[User]: {content}")
elif role == "assistant":
print(f"[Assistant]: {content}")
if msg.get("tool_calls"):
for tc in msg["tool_calls"]:
func = tc.get("function", {})
print(f" -> {func.get('name')}({func.get('arguments', '{}')[:50]})")
elif role == "tool":
print(f" [Result]: {content[:100]}...")
return 0
elif parsed.command == "delete":
if manager.delete_template(parsed.template_id):
print(f"Deleted template: {parsed.template_id}")
return 0
else:
print(f"Template {parsed.template_id} not found")
return 1
return 1
if __name__ == "__main__":
import sys
sys.exit(warm_session_cli(sys.argv[1:]))