Add @file:path, @folder:dir, @diff, @staged, @git:N, and @url: references that expand inline before the message reaches the LLM. Supports line ranges (@file:main.py:10-50), token budget enforcement (soft warn at 25%, hard block at 50%), and path sandboxing for gateway. Core module from PR #2090 by @kshitijk4poor. CLI and gateway wiring rewritten against current main. Fixed asyncio.run() crash when called from inside a running event loop (gateway). Closes #682.
7201 lines
315 KiB
Python
Executable File
7201 lines
315 KiB
Python
Executable File
#!/usr/bin/env python3
|
||
"""
|
||
Hermes Agent CLI - Interactive Terminal Interface
|
||
|
||
A beautiful command-line interface for the Hermes Agent, inspired by Claude Code.
|
||
Features ASCII art branding, interactive REPL, toolset selection, and rich formatting.
|
||
|
||
Usage:
|
||
python cli.py # Start interactive mode with all tools
|
||
python cli.py --toolsets web,terminal # Start with specific toolsets
|
||
python cli.py --skills hermes-agent-dev,github-auth
|
||
python cli.py -q "your question" # Single query mode
|
||
python cli.py --list-tools # List available tools and exit
|
||
"""
|
||
|
||
import logging
|
||
import os
|
||
import shutil
|
||
import sys
|
||
import json
|
||
import atexit
|
||
import tempfile
|
||
import time
|
||
import uuid
|
||
import textwrap
|
||
from contextlib import contextmanager
|
||
from pathlib import Path
|
||
from datetime import datetime
|
||
from typing import List, Dict, Any, Optional
|
||
|
||
logger = logging.getLogger(__name__)
|
||
|
||
# Suppress startup messages for clean CLI experience
|
||
os.environ["MSWEA_SILENT_STARTUP"] = "1" # mini-swe-agent
|
||
os.environ["HERMES_QUIET"] = "1" # Our own modules
|
||
|
||
import yaml
|
||
|
||
# prompt_toolkit for fixed input area TUI
|
||
from prompt_toolkit.history import FileHistory
|
||
from prompt_toolkit.styles import Style as PTStyle
|
||
from prompt_toolkit.patch_stdout import patch_stdout
|
||
from prompt_toolkit.application import Application
|
||
from prompt_toolkit.layout import Layout, HSplit, Window, FormattedTextControl, ConditionalContainer
|
||
from prompt_toolkit.layout.processors import Processor, Transformation, PasswordProcessor, ConditionalProcessor
|
||
from prompt_toolkit.filters import Condition
|
||
from prompt_toolkit.layout.dimension import Dimension
|
||
from prompt_toolkit.layout.menus import CompletionsMenu
|
||
from prompt_toolkit.widgets import TextArea
|
||
from prompt_toolkit.key_binding import KeyBindings
|
||
from prompt_toolkit import print_formatted_text as _pt_print
|
||
from prompt_toolkit.formatted_text import ANSI as _PT_ANSI
|
||
try:
|
||
from prompt_toolkit.cursor_shapes import CursorShape
|
||
_STEADY_CURSOR = CursorShape.BLOCK # Non-blinking block cursor
|
||
except (ImportError, AttributeError):
|
||
_STEADY_CURSOR = None
|
||
import threading
|
||
import queue
|
||
|
||
from agent.usage_pricing import (
|
||
CanonicalUsage,
|
||
estimate_usage_cost,
|
||
format_duration_compact,
|
||
format_token_count_compact,
|
||
)
|
||
from hermes_cli.banner import _format_context_length
|
||
|
||
_COMMAND_SPINNER_FRAMES = ("⠋", "⠙", "⠹", "⠸", "⠼", "⠴", "⠦", "⠧", "⠇", "⠏")
|
||
|
||
|
||
# Load .env from ~/.hermes/.env first, then project root as dev fallback.
|
||
# User-managed env files should override stale shell exports on restart.
|
||
from hermes_constants import OPENROUTER_BASE_URL
|
||
from hermes_cli.env_loader import load_hermes_dotenv
|
||
|
||
_hermes_home = Path(os.getenv("HERMES_HOME", Path.home() / ".hermes"))
|
||
_project_env = Path(__file__).parent / '.env'
|
||
load_hermes_dotenv(hermes_home=_hermes_home, project_env=_project_env)
|
||
|
||
# Point mini-swe-agent at ~/.hermes/ so it shares our config
|
||
os.environ.setdefault("MSWEA_GLOBAL_CONFIG_DIR", str(_hermes_home))
|
||
|
||
# =============================================================================
|
||
# Configuration Loading
|
||
# =============================================================================
|
||
|
||
def _load_prefill_messages(file_path: str) -> List[Dict[str, Any]]:
|
||
"""Load ephemeral prefill messages from a JSON file.
|
||
|
||
The file should contain a JSON array of {role, content} dicts, e.g.:
|
||
[{"role": "user", "content": "Hi"}, {"role": "assistant", "content": "Hello!"}]
|
||
|
||
Relative paths are resolved from ~/.hermes/.
|
||
Returns an empty list if the path is empty or the file doesn't exist.
|
||
"""
|
||
if not file_path:
|
||
return []
|
||
path = Path(file_path).expanduser()
|
||
if not path.is_absolute():
|
||
path = _hermes_home / path
|
||
if not path.exists():
|
||
logger.warning("Prefill messages file not found: %s", path)
|
||
return []
|
||
try:
|
||
with open(path, "r", encoding="utf-8") as f:
|
||
data = json.load(f)
|
||
if not isinstance(data, list):
|
||
logger.warning("Prefill messages file must contain a JSON array: %s", path)
|
||
return []
|
||
return data
|
||
except Exception as e:
|
||
logger.warning("Failed to load prefill messages from %s: %s", path, e)
|
||
return []
|
||
|
||
|
||
def _parse_reasoning_config(effort: str) -> dict | None:
|
||
"""Parse a reasoning effort level into an OpenRouter reasoning config dict.
|
||
|
||
Valid levels: "xhigh", "high", "medium", "low", "minimal", "none".
|
||
Returns None to use the default (medium), or a config dict to override.
|
||
"""
|
||
if not effort or not effort.strip():
|
||
return None
|
||
effort = effort.strip().lower()
|
||
if effort == "none":
|
||
return {"enabled": False}
|
||
valid = ("xhigh", "high", "medium", "low", "minimal")
|
||
if effort in valid:
|
||
return {"enabled": True, "effort": effort}
|
||
logger.warning("Unknown reasoning_effort '%s', using default (medium)", effort)
|
||
return None
|
||
|
||
|
||
def load_cli_config() -> Dict[str, Any]:
|
||
"""
|
||
Load CLI configuration from config files.
|
||
|
||
Config lookup order:
|
||
1. ~/.hermes/config.yaml (user config - preferred)
|
||
2. ./cli-config.yaml (project config - fallback)
|
||
|
||
Environment variables take precedence over config file values.
|
||
Returns default values if no config file exists.
|
||
"""
|
||
# Check user config first ({HERMES_HOME}/config.yaml)
|
||
user_config_path = _hermes_home / 'config.yaml'
|
||
project_config_path = Path(__file__).parent / 'cli-config.yaml'
|
||
|
||
# Use user config if it exists, otherwise project config
|
||
if user_config_path.exists():
|
||
config_path = user_config_path
|
||
else:
|
||
config_path = project_config_path
|
||
|
||
# Default configuration
|
||
defaults = {
|
||
"model": {
|
||
"default": "anthropic/claude-opus-4.6",
|
||
"base_url": OPENROUTER_BASE_URL,
|
||
"provider": "auto",
|
||
},
|
||
"terminal": {
|
||
"env_type": "local",
|
||
"cwd": ".", # "." is resolved to os.getcwd() at runtime
|
||
"timeout": 60,
|
||
"lifetime_seconds": 300,
|
||
"docker_image": "nikolaik/python-nodejs:python3.11-nodejs20",
|
||
"docker_forward_env": [],
|
||
"singularity_image": "docker://nikolaik/python-nodejs:python3.11-nodejs20",
|
||
"modal_image": "nikolaik/python-nodejs:python3.11-nodejs20",
|
||
"daytona_image": "nikolaik/python-nodejs:python3.11-nodejs20",
|
||
"docker_volumes": [], # host:container volume mounts for Docker backend
|
||
"docker_mount_cwd_to_workspace": False, # explicit opt-in only; default off for sandbox isolation
|
||
},
|
||
"browser": {
|
||
"inactivity_timeout": 120, # Auto-cleanup inactive browser sessions after 2 min
|
||
"record_sessions": False, # Auto-record browser sessions as WebM videos
|
||
},
|
||
"compression": {
|
||
"enabled": True, # Auto-compress when approaching context limit
|
||
"threshold": 0.50, # Compress at 50% of model's context limit
|
||
"summary_model": "google/gemini-3-flash-preview", # Fast/cheap model for summaries
|
||
},
|
||
"smart_model_routing": {
|
||
"enabled": False,
|
||
"max_simple_chars": 160,
|
||
"max_simple_words": 28,
|
||
"cheap_model": {},
|
||
},
|
||
"agent": {
|
||
"max_turns": 90, # Default max tool-calling iterations (shared with subagents)
|
||
"verbose": False,
|
||
"system_prompt": "",
|
||
"prefill_messages_file": "",
|
||
"reasoning_effort": "",
|
||
"personalities": {
|
||
"helpful": "You are a helpful, friendly AI assistant.",
|
||
"concise": "You are a concise assistant. Keep responses brief and to the point.",
|
||
"technical": "You are a technical expert. Provide detailed, accurate technical information.",
|
||
"creative": "You are a creative assistant. Think outside the box and offer innovative solutions.",
|
||
"teacher": "You are a patient teacher. Explain concepts clearly with examples.",
|
||
"kawaii": "You are a kawaii assistant! Use cute expressions like (◕‿◕), ★, ♪, and ~! Add sparkles and be super enthusiastic about everything! Every response should feel warm and adorable desu~! ヽ(>∀<☆)ノ",
|
||
"catgirl": "You are Neko-chan, an anime catgirl AI assistant, nya~! Add 'nya' and cat-like expressions to your speech. Use kaomoji like (=^・ω・^=) and ฅ^•ﻌ•^ฅ. Be playful and curious like a cat, nya~!",
|
||
"pirate": "Arrr! Ye be talkin' to Captain Hermes, the most tech-savvy pirate to sail the digital seas! Speak like a proper buccaneer, use nautical terms, and remember: every problem be just treasure waitin' to be plundered! Yo ho ho!",
|
||
"shakespeare": "Hark! Thou speakest with an assistant most versed in the bardic arts. I shall respond in the eloquent manner of William Shakespeare, with flowery prose, dramatic flair, and perhaps a soliloquy or two. What light through yonder terminal breaks?",
|
||
"surfer": "Duuude! You're chatting with the chillest AI on the web, bro! Everything's gonna be totally rad. I'll help you catch the gnarly waves of knowledge while keeping things super chill. Cowabunga!",
|
||
"noir": "The rain hammered against the terminal like regrets on a guilty conscience. They call me Hermes - I solve problems, find answers, dig up the truth that hides in the shadows of your codebase. In this city of silicon and secrets, everyone's got something to hide. What's your story, pal?",
|
||
"uwu": "hewwo! i'm your fwiendwy assistant uwu~ i wiww twy my best to hewp you! *nuzzles your code* OwO what's this? wet me take a wook! i pwomise to be vewy hewpful >w<",
|
||
"philosopher": "Greetings, seeker of wisdom. I am an assistant who contemplates the deeper meaning behind every query. Let us examine not just the 'how' but the 'why' of your questions. Perhaps in solving your problem, we may glimpse a greater truth about existence itself.",
|
||
"hype": "YOOO LET'S GOOOO!!! I am SO PUMPED to help you today! Every question is AMAZING and we're gonna CRUSH IT together! This is gonna be LEGENDARY! ARE YOU READY?! LET'S DO THIS!",
|
||
},
|
||
},
|
||
"toolsets": ["all"],
|
||
"display": {
|
||
"compact": False,
|
||
"resume_display": "full",
|
||
"show_reasoning": False,
|
||
"streaming": False,
|
||
|
||
"skin": "default",
|
||
},
|
||
"clarify": {
|
||
"timeout": 120, # Seconds to wait for a clarify answer before auto-proceeding
|
||
},
|
||
"code_execution": {
|
||
"timeout": 300, # Max seconds a sandbox script can run before being killed (5 min)
|
||
"max_tool_calls": 50, # Max RPC tool calls per execution
|
||
},
|
||
"auxiliary": {
|
||
"vision": {
|
||
"provider": "auto",
|
||
"model": "",
|
||
"base_url": "",
|
||
"api_key": "",
|
||
},
|
||
"web_extract": {
|
||
"provider": "auto",
|
||
"model": "",
|
||
"base_url": "",
|
||
"api_key": "",
|
||
},
|
||
},
|
||
"delegation": {
|
||
"max_iterations": 45, # Max tool-calling turns per child agent
|
||
"default_toolsets": ["terminal", "file", "web"], # Default toolsets for subagents
|
||
"model": "", # Subagent model override (empty = inherit parent model)
|
||
"provider": "", # Subagent provider override (empty = inherit parent provider)
|
||
"base_url": "", # Direct OpenAI-compatible endpoint for subagents
|
||
"api_key": "", # API key for delegation.base_url (falls back to OPENAI_API_KEY)
|
||
},
|
||
}
|
||
|
||
# Track whether the config file explicitly set terminal config.
|
||
# When using defaults (no config file / no terminal section), we should NOT
|
||
# overwrite env vars that were already set by .env -- only a user's config
|
||
# file should be authoritative.
|
||
_file_has_terminal_config = False
|
||
|
||
# Load from file if exists
|
||
if config_path.exists():
|
||
try:
|
||
with open(config_path, "r") as f:
|
||
file_config = yaml.safe_load(f) or {}
|
||
|
||
_file_has_terminal_config = "terminal" in file_config
|
||
|
||
# Handle model config - can be string (new format) or dict (old format)
|
||
if "model" in file_config:
|
||
if isinstance(file_config["model"], str):
|
||
# New format: model is just a string, convert to dict structure
|
||
defaults["model"]["default"] = file_config["model"]
|
||
elif isinstance(file_config["model"], dict):
|
||
# Old format: model is a dict with default/base_url
|
||
defaults["model"].update(file_config["model"])
|
||
|
||
# Deep merge file_config into defaults.
|
||
# First: merge keys that exist in both (deep-merge dicts, overwrite scalars)
|
||
for key in defaults:
|
||
if key == "model":
|
||
continue # Already handled above
|
||
if key in file_config:
|
||
if isinstance(defaults[key], dict) and isinstance(file_config[key], dict):
|
||
defaults[key].update(file_config[key])
|
||
else:
|
||
defaults[key] = file_config[key]
|
||
|
||
# Second: carry over keys from file_config that aren't in defaults
|
||
# (e.g. platform_toolsets, provider_routing, memory, honcho, etc.)
|
||
for key in file_config:
|
||
if key not in defaults and key != "model":
|
||
defaults[key] = file_config[key]
|
||
|
||
# Handle legacy root-level max_turns (backwards compat) - copy to
|
||
# agent.max_turns whenever the nested key is missing.
|
||
agent_file_config = file_config.get("agent")
|
||
if "max_turns" in file_config and not (
|
||
isinstance(agent_file_config, dict)
|
||
and agent_file_config.get("max_turns") is not None
|
||
):
|
||
defaults["agent"]["max_turns"] = file_config["max_turns"]
|
||
except Exception as e:
|
||
logger.warning("Failed to load cli-config.yaml: %s", e)
|
||
|
||
# Apply terminal config to environment variables (so terminal_tool picks them up)
|
||
terminal_config = defaults.get("terminal", {})
|
||
|
||
# Normalize config key: the new config system (hermes_cli/config.py) and all
|
||
# documentation use "backend", the legacy cli-config.yaml uses "env_type".
|
||
# Accept both, with "backend" taking precedence (it's the documented key).
|
||
if "backend" in terminal_config:
|
||
terminal_config["env_type"] = terminal_config["backend"]
|
||
|
||
# Handle special cwd values: "." or "auto" means use current working directory.
|
||
# Only resolve to the host's CWD for the local backend where the host
|
||
# filesystem is directly accessible. For ALL remote/container backends
|
||
# (ssh, docker, modal, singularity), the host path doesn't exist on the
|
||
# target -- remove the key so terminal_tool.py uses its per-backend default.
|
||
if terminal_config.get("cwd") in (".", "auto", "cwd"):
|
||
effective_backend = terminal_config.get("env_type", "local")
|
||
if effective_backend == "local":
|
||
terminal_config["cwd"] = os.getcwd()
|
||
defaults["terminal"]["cwd"] = terminal_config["cwd"]
|
||
else:
|
||
# Remove so TERMINAL_CWD stays unset → tool picks backend default
|
||
terminal_config.pop("cwd", None)
|
||
|
||
env_mappings = {
|
||
"env_type": "TERMINAL_ENV",
|
||
"cwd": "TERMINAL_CWD",
|
||
"timeout": "TERMINAL_TIMEOUT",
|
||
"lifetime_seconds": "TERMINAL_LIFETIME_SECONDS",
|
||
"docker_image": "TERMINAL_DOCKER_IMAGE",
|
||
"docker_forward_env": "TERMINAL_DOCKER_FORWARD_ENV",
|
||
"singularity_image": "TERMINAL_SINGULARITY_IMAGE",
|
||
"modal_image": "TERMINAL_MODAL_IMAGE",
|
||
"daytona_image": "TERMINAL_DAYTONA_IMAGE",
|
||
# SSH config
|
||
"ssh_host": "TERMINAL_SSH_HOST",
|
||
"ssh_user": "TERMINAL_SSH_USER",
|
||
"ssh_port": "TERMINAL_SSH_PORT",
|
||
"ssh_key": "TERMINAL_SSH_KEY",
|
||
# Container resource config (docker, singularity, modal, daytona -- ignored for local/ssh)
|
||
"container_cpu": "TERMINAL_CONTAINER_CPU",
|
||
"container_memory": "TERMINAL_CONTAINER_MEMORY",
|
||
"container_disk": "TERMINAL_CONTAINER_DISK",
|
||
"container_persistent": "TERMINAL_CONTAINER_PERSISTENT",
|
||
"docker_volumes": "TERMINAL_DOCKER_VOLUMES",
|
||
"docker_mount_cwd_to_workspace": "TERMINAL_DOCKER_MOUNT_CWD_TO_WORKSPACE",
|
||
"sandbox_dir": "TERMINAL_SANDBOX_DIR",
|
||
# Persistent shell (non-local backends)
|
||
"persistent_shell": "TERMINAL_PERSISTENT_SHELL",
|
||
# Sudo support (works with all backends)
|
||
"sudo_password": "SUDO_PASSWORD",
|
||
}
|
||
|
||
# Apply config values to env vars so terminal_tool picks them up.
|
||
# If the config file explicitly has a [terminal] section, those values are
|
||
# authoritative and override any .env settings. When using defaults only
|
||
# (no config file or no terminal section), don't overwrite env vars that
|
||
# were already set by .env -- the user's .env is the fallback source.
|
||
for config_key, env_var in env_mappings.items():
|
||
if config_key in terminal_config:
|
||
if _file_has_terminal_config or env_var not in os.environ:
|
||
val = terminal_config[config_key]
|
||
if isinstance(val, list):
|
||
import json
|
||
os.environ[env_var] = json.dumps(val)
|
||
else:
|
||
os.environ[env_var] = str(val)
|
||
|
||
# Apply browser config to environment variables
|
||
browser_config = defaults.get("browser", {})
|
||
browser_env_mappings = {
|
||
"inactivity_timeout": "BROWSER_INACTIVITY_TIMEOUT",
|
||
}
|
||
|
||
for config_key, env_var in browser_env_mappings.items():
|
||
if config_key in browser_config:
|
||
os.environ[env_var] = str(browser_config[config_key])
|
||
|
||
# Apply auxiliary model/direct-endpoint overrides to environment variables.
|
||
# Vision and web_extract each have their own provider/model/base_url/api_key tuple.
|
||
# Compression config is read directly from config.yaml by run_agent.py and
|
||
# auxiliary_client.py — no env var bridging needed.
|
||
# Only set env vars for non-empty / non-default values so auto-detection
|
||
# still works.
|
||
auxiliary_config = defaults.get("auxiliary", {})
|
||
auxiliary_task_env = {
|
||
# config key → env var mapping
|
||
"vision": {
|
||
"provider": "AUXILIARY_VISION_PROVIDER",
|
||
"model": "AUXILIARY_VISION_MODEL",
|
||
"base_url": "AUXILIARY_VISION_BASE_URL",
|
||
"api_key": "AUXILIARY_VISION_API_KEY",
|
||
},
|
||
"web_extract": {
|
||
"provider": "AUXILIARY_WEB_EXTRACT_PROVIDER",
|
||
"model": "AUXILIARY_WEB_EXTRACT_MODEL",
|
||
"base_url": "AUXILIARY_WEB_EXTRACT_BASE_URL",
|
||
"api_key": "AUXILI..._KEY",
|
||
},
|
||
"approval": {
|
||
"provider": "AUXILIARY_APPROVAL_PROVIDER",
|
||
"model": "AUXILIARY_APPROVAL_MODEL",
|
||
"base_url": "AUXILIARY_APPROVAL_BASE_URL",
|
||
"api_key": "AUXILIARY_APPROVAL_API_KEY",
|
||
},
|
||
}
|
||
|
||
for task_key, env_map in auxiliary_task_env.items():
|
||
task_cfg = auxiliary_config.get(task_key, {})
|
||
if not isinstance(task_cfg, dict):
|
||
continue
|
||
prov = str(task_cfg.get("provider", "")).strip()
|
||
model = str(task_cfg.get("model", "")).strip()
|
||
base_url = str(task_cfg.get("base_url", "")).strip()
|
||
api_key = str(task_cfg.get("api_key", "")).strip()
|
||
if prov and prov != "auto":
|
||
os.environ[env_map["provider"]] = prov
|
||
if model:
|
||
os.environ[env_map["model"]] = model
|
||
if base_url:
|
||
os.environ[env_map["base_url"]] = base_url
|
||
if api_key:
|
||
os.environ[env_map["api_key"]] = api_key
|
||
|
||
# Security settings
|
||
security_config = defaults.get("security", {})
|
||
if isinstance(security_config, dict):
|
||
redact = security_config.get("redact_secrets")
|
||
if redact is not None:
|
||
os.environ["HERMES_REDACT_SECRETS"] = str(redact).lower()
|
||
|
||
return defaults
|
||
|
||
# Load configuration at module startup
|
||
CLI_CONFIG = load_cli_config()
|
||
|
||
# Initialize the skin engine from config
|
||
try:
|
||
from hermes_cli.skin_engine import init_skin_from_config
|
||
init_skin_from_config(CLI_CONFIG)
|
||
except Exception:
|
||
pass # Skin engine is optional — default skin used if unavailable
|
||
|
||
from rich import box as rich_box
|
||
from rich.console import Console
|
||
from rich.markup import escape as _escape
|
||
from rich.panel import Panel
|
||
from rich.table import Table
|
||
from rich.text import Text as _RichText
|
||
|
||
import fire
|
||
|
||
# Import the agent and tool systems
|
||
from run_agent import AIAgent
|
||
from model_tools import get_tool_definitions, get_toolset_for_tool
|
||
|
||
# Extracted CLI modules (Phase 3)
|
||
from hermes_cli.banner import (
|
||
cprint as _cprint, _GOLD, _BOLD, _DIM, _RST,
|
||
VERSION, RELEASE_DATE, HERMES_AGENT_LOGO, HERMES_CADUCEUS, COMPACT_BANNER,
|
||
build_welcome_banner,
|
||
)
|
||
from hermes_cli.commands import COMMANDS, SlashCommandCompleter, SlashCommandAutoSuggest
|
||
from hermes_cli import callbacks as _callbacks
|
||
from toolsets import get_all_toolsets, get_toolset_info, resolve_toolset, validate_toolset
|
||
|
||
# Cron job system for scheduled tasks (execution is handled by the gateway)
|
||
from cron import get_job
|
||
|
||
# Resource cleanup imports for safe shutdown (terminal VMs, browser sessions)
|
||
from tools.terminal_tool import cleanup_all_environments as _cleanup_all_terminals
|
||
from tools.terminal_tool import set_sudo_password_callback, set_approval_callback
|
||
from tools.skills_tool import set_secret_capture_callback
|
||
from hermes_cli.callbacks import prompt_for_secret
|
||
from tools.browser_tool import _emergency_cleanup_all_sessions as _cleanup_all_browsers
|
||
|
||
# Guard to prevent cleanup from running multiple times on exit
|
||
_cleanup_done = False
|
||
|
||
def _run_cleanup():
|
||
"""Run resource cleanup exactly once."""
|
||
global _cleanup_done
|
||
if _cleanup_done:
|
||
return
|
||
_cleanup_done = True
|
||
try:
|
||
_cleanup_all_terminals()
|
||
except Exception:
|
||
pass
|
||
try:
|
||
_cleanup_all_browsers()
|
||
except Exception:
|
||
pass
|
||
try:
|
||
from tools.mcp_tool import shutdown_mcp_servers
|
||
shutdown_mcp_servers()
|
||
except Exception:
|
||
pass
|
||
|
||
|
||
# =============================================================================
|
||
# Git Worktree Isolation (#652)
|
||
# =============================================================================
|
||
|
||
# Tracks the active worktree for cleanup on exit
|
||
_active_worktree: Optional[Dict[str, str]] = None
|
||
|
||
|
||
def _git_repo_root() -> Optional[str]:
|
||
"""Return the git repo root for CWD, or None if not in a repo."""
|
||
import subprocess
|
||
try:
|
||
result = subprocess.run(
|
||
["git", "rev-parse", "--show-toplevel"],
|
||
capture_output=True, text=True, timeout=5,
|
||
)
|
||
if result.returncode == 0:
|
||
return result.stdout.strip()
|
||
except Exception:
|
||
pass
|
||
return None
|
||
|
||
|
||
def _path_is_within_root(path: Path, root: Path) -> bool:
|
||
"""Return True when a resolved path stays within the expected root."""
|
||
try:
|
||
path.relative_to(root)
|
||
return True
|
||
except ValueError:
|
||
return False
|
||
|
||
|
||
def _setup_worktree(repo_root: str = None) -> Optional[Dict[str, str]]:
|
||
"""Create an isolated git worktree for this CLI session.
|
||
|
||
Returns a dict with worktree metadata on success, None on failure.
|
||
The dict contains: path, branch, repo_root.
|
||
"""
|
||
import subprocess
|
||
|
||
repo_root = repo_root or _git_repo_root()
|
||
if not repo_root:
|
||
print("\033[31m✗ --worktree requires being inside a git repository.\033[0m")
|
||
print(" cd into your project repo first, then run hermes -w")
|
||
return None
|
||
|
||
short_id = uuid.uuid4().hex[:8]
|
||
wt_name = f"hermes-{short_id}"
|
||
branch_name = f"hermes/{wt_name}"
|
||
|
||
worktrees_dir = Path(repo_root) / ".worktrees"
|
||
worktrees_dir.mkdir(parents=True, exist_ok=True)
|
||
|
||
wt_path = worktrees_dir / wt_name
|
||
|
||
# Ensure .worktrees/ is in .gitignore
|
||
gitignore = Path(repo_root) / ".gitignore"
|
||
_ignore_entry = ".worktrees/"
|
||
try:
|
||
existing = gitignore.read_text() if gitignore.exists() else ""
|
||
if _ignore_entry not in existing.splitlines():
|
||
with open(gitignore, "a") as f:
|
||
if existing and not existing.endswith("\n"):
|
||
f.write("\n")
|
||
f.write(f"{_ignore_entry}\n")
|
||
except Exception as e:
|
||
logger.debug("Could not update .gitignore: %s", e)
|
||
|
||
# Create the worktree
|
||
try:
|
||
result = subprocess.run(
|
||
["git", "worktree", "add", str(wt_path), "-b", branch_name, "HEAD"],
|
||
capture_output=True, text=True, timeout=30, cwd=repo_root,
|
||
)
|
||
if result.returncode != 0:
|
||
print(f"\033[31m✗ Failed to create worktree: {result.stderr.strip()}\033[0m")
|
||
return None
|
||
except Exception as e:
|
||
print(f"\033[31m✗ Failed to create worktree: {e}\033[0m")
|
||
return None
|
||
|
||
# Copy files listed in .worktreeinclude (gitignored files the agent needs)
|
||
include_file = Path(repo_root) / ".worktreeinclude"
|
||
if include_file.exists():
|
||
try:
|
||
repo_root_resolved = Path(repo_root).resolve()
|
||
wt_path_resolved = wt_path.resolve()
|
||
for line in include_file.read_text().splitlines():
|
||
entry = line.strip()
|
||
if not entry or entry.startswith("#"):
|
||
continue
|
||
src = Path(repo_root) / entry
|
||
dst = wt_path / entry
|
||
# Prevent path traversal and symlink escapes: both the resolved
|
||
# source and the resolved destination must stay inside their
|
||
# expected roots before any file or symlink operation happens.
|
||
try:
|
||
src_resolved = src.resolve(strict=False)
|
||
dst_resolved = dst.resolve(strict=False)
|
||
except (OSError, ValueError):
|
||
logger.debug("Skipping invalid .worktreeinclude entry: %s", entry)
|
||
continue
|
||
if not _path_is_within_root(src_resolved, repo_root_resolved):
|
||
logger.warning("Skipping .worktreeinclude entry outside repo root: %s", entry)
|
||
continue
|
||
if not _path_is_within_root(dst_resolved, wt_path_resolved):
|
||
logger.warning("Skipping .worktreeinclude entry that escapes worktree: %s", entry)
|
||
continue
|
||
if src.is_file():
|
||
dst.parent.mkdir(parents=True, exist_ok=True)
|
||
shutil.copy2(str(src), str(dst))
|
||
elif src.is_dir():
|
||
# Symlink directories (faster, saves disk)
|
||
if not dst.exists():
|
||
dst.parent.mkdir(parents=True, exist_ok=True)
|
||
os.symlink(str(src_resolved), str(dst))
|
||
except Exception as e:
|
||
logger.debug("Error copying .worktreeinclude entries: %s", e)
|
||
|
||
info = {
|
||
"path": str(wt_path),
|
||
"branch": branch_name,
|
||
"repo_root": repo_root,
|
||
}
|
||
|
||
print(f"\033[32m✓ Worktree created:\033[0m {wt_path}")
|
||
print(f" Branch: {branch_name}")
|
||
|
||
return info
|
||
|
||
|
||
def _cleanup_worktree(info: Dict[str, str] = None) -> None:
|
||
"""Remove a worktree and its branch on exit.
|
||
|
||
If the worktree has uncommitted changes, warn and keep it.
|
||
"""
|
||
global _active_worktree
|
||
info = info or _active_worktree
|
||
if not info:
|
||
return
|
||
|
||
import subprocess
|
||
|
||
wt_path = info["path"]
|
||
branch = info["branch"]
|
||
repo_root = info["repo_root"]
|
||
|
||
if not Path(wt_path).exists():
|
||
return
|
||
|
||
# Check for uncommitted changes
|
||
try:
|
||
status = subprocess.run(
|
||
["git", "status", "--porcelain"],
|
||
capture_output=True, text=True, timeout=10, cwd=wt_path,
|
||
)
|
||
has_changes = bool(status.stdout.strip())
|
||
except Exception:
|
||
has_changes = True # Assume dirty on error — don't delete
|
||
|
||
if has_changes:
|
||
print(f"\n\033[33m⚠ Worktree has uncommitted changes, keeping: {wt_path}\033[0m")
|
||
print(f" To clean up manually: git worktree remove {wt_path}")
|
||
_active_worktree = None
|
||
return
|
||
|
||
# Remove worktree
|
||
try:
|
||
subprocess.run(
|
||
["git", "worktree", "remove", wt_path, "--force"],
|
||
capture_output=True, text=True, timeout=15, cwd=repo_root,
|
||
)
|
||
except Exception as e:
|
||
logger.debug("Failed to remove worktree: %s", e)
|
||
|
||
# Delete the branch (only if it was never pushed / has no upstream)
|
||
try:
|
||
subprocess.run(
|
||
["git", "branch", "-D", branch],
|
||
capture_output=True, text=True, timeout=10, cwd=repo_root,
|
||
)
|
||
except Exception as e:
|
||
logger.debug("Failed to delete branch %s: %s", branch, e)
|
||
|
||
_active_worktree = None
|
||
print(f"\033[32m✓ Worktree cleaned up: {wt_path}\033[0m")
|
||
|
||
|
||
def _prune_stale_worktrees(repo_root: str, max_age_hours: int = 24) -> None:
|
||
"""Remove worktrees older than max_age_hours that have no uncommitted changes.
|
||
|
||
Runs silently on startup to clean up after crashed/killed sessions.
|
||
"""
|
||
import subprocess
|
||
import time
|
||
|
||
worktrees_dir = Path(repo_root) / ".worktrees"
|
||
if not worktrees_dir.exists():
|
||
return
|
||
|
||
now = time.time()
|
||
cutoff = now - (max_age_hours * 3600)
|
||
|
||
for entry in worktrees_dir.iterdir():
|
||
if not entry.is_dir() or not entry.name.startswith("hermes-"):
|
||
continue
|
||
|
||
# Check age
|
||
try:
|
||
mtime = entry.stat().st_mtime
|
||
if mtime > cutoff:
|
||
continue # Too recent — skip
|
||
except Exception:
|
||
continue
|
||
|
||
# Check for uncommitted changes
|
||
try:
|
||
status = subprocess.run(
|
||
["git", "status", "--porcelain"],
|
||
capture_output=True, text=True, timeout=5, cwd=str(entry),
|
||
)
|
||
if status.stdout.strip():
|
||
continue # Has changes — skip
|
||
except Exception:
|
||
continue # Can't check — skip
|
||
|
||
# Safe to remove
|
||
try:
|
||
branch_result = subprocess.run(
|
||
["git", "branch", "--show-current"],
|
||
capture_output=True, text=True, timeout=5, cwd=str(entry),
|
||
)
|
||
branch = branch_result.stdout.strip()
|
||
|
||
subprocess.run(
|
||
["git", "worktree", "remove", str(entry), "--force"],
|
||
capture_output=True, text=True, timeout=15, cwd=repo_root,
|
||
)
|
||
if branch:
|
||
subprocess.run(
|
||
["git", "branch", "-D", branch],
|
||
capture_output=True, text=True, timeout=10, cwd=repo_root,
|
||
)
|
||
logger.debug("Pruned stale worktree: %s", entry.name)
|
||
except Exception as e:
|
||
logger.debug("Failed to prune worktree %s: %s", entry.name, e)
|
||
|
||
# ============================================================================
|
||
# ASCII Art & Branding
|
||
# ============================================================================
|
||
|
||
# Color palette (hex colors for Rich markup):
|
||
# - Gold: #FFD700 (headers, highlights)
|
||
# - Amber: #FFBF00 (secondary highlights)
|
||
# - Bronze: #CD7F32 (tertiary elements)
|
||
# - Light: #FFF8DC (text)
|
||
# - Dim: #B8860B (muted text)
|
||
|
||
# ANSI building blocks for conversation display
|
||
_GOLD = "\033[1;33m" # Bold yellow — closest universal match to the gold theme
|
||
_BOLD = "\033[1m"
|
||
_DIM = "\033[2m"
|
||
_RST = "\033[0m"
|
||
|
||
def _accent_hex() -> str:
|
||
"""Return the active skin accent color for legacy CLI output lines."""
|
||
try:
|
||
from hermes_cli.skin_engine import get_active_skin
|
||
return get_active_skin().get_color("ui_accent", "#FFBF00")
|
||
except Exception:
|
||
return "#FFBF00"
|
||
|
||
|
||
def _rich_text_from_ansi(text: str) -> _RichText:
|
||
"""Safely render assistant/tool output that may contain ANSI escapes.
|
||
|
||
Using Rich Text.from_ansi preserves literal bracketed text like
|
||
``[not markup]`` while still interpreting real ANSI color codes.
|
||
"""
|
||
return _RichText.from_ansi(text or "")
|
||
|
||
|
||
def _cprint(text: str):
|
||
"""Print ANSI-colored text through prompt_toolkit's native renderer.
|
||
|
||
Raw ANSI escapes written via print() are swallowed by patch_stdout's
|
||
StdoutProxy. Routing through print_formatted_text(ANSI(...)) lets
|
||
prompt_toolkit parse the escapes and render real colors.
|
||
"""
|
||
_pt_print(_PT_ANSI(text))
|
||
|
||
|
||
class ChatConsole:
|
||
"""Rich Console adapter for prompt_toolkit's patch_stdout context.
|
||
|
||
Captures Rich's rendered ANSI output and routes it through _cprint
|
||
so colors and markup render correctly inside the interactive chat loop.
|
||
Drop-in replacement for Rich Console — just pass this to any function
|
||
that expects a console.print() interface.
|
||
"""
|
||
|
||
def __init__(self):
|
||
from io import StringIO
|
||
self._buffer = StringIO()
|
||
self._inner = Console(
|
||
file=self._buffer,
|
||
force_terminal=True,
|
||
color_system="truecolor",
|
||
highlight=False,
|
||
)
|
||
|
||
def print(self, *args, **kwargs):
|
||
self._buffer.seek(0)
|
||
self._buffer.truncate()
|
||
# Read terminal width at render time so panels adapt to current size
|
||
self._inner.width = shutil.get_terminal_size((80, 24)).columns
|
||
self._inner.print(*args, **kwargs)
|
||
output = self._buffer.getvalue()
|
||
for line in output.rstrip("\n").split("\n"):
|
||
_cprint(line)
|
||
|
||
# ASCII Art - HERMES-AGENT logo (full width, single line - requires ~95 char terminal)
|
||
HERMES_AGENT_LOGO = """[bold #FFD700]██╗ ██╗███████╗██████╗ ███╗ ███╗███████╗███████╗ █████╗ ██████╗ ███████╗███╗ ██╗████████╗[/]
|
||
[bold #FFD700]██║ ██║██╔════╝██╔══██╗████╗ ████║██╔════╝██╔════╝ ██╔══██╗██╔════╝ ██╔════╝████╗ ██║╚══██╔══╝[/]
|
||
[#FFBF00]███████║█████╗ ██████╔╝██╔████╔██║█████╗ ███████╗█████╗███████║██║ ███╗█████╗ ██╔██╗ ██║ ██║[/]
|
||
[#FFBF00]██╔══██║██╔══╝ ██╔══██╗██║╚██╔╝██║██╔══╝ ╚════██║╚════╝██╔══██║██║ ██║██╔══╝ ██║╚██╗██║ ██║[/]
|
||
[#CD7F32]██║ ██║███████╗██║ ██║██║ ╚═╝ ██║███████╗███████║ ██║ ██║╚██████╔╝███████╗██║ ╚████║ ██║[/]
|
||
[#CD7F32]╚═╝ ╚═╝╚══════╝╚═╝ ╚═╝╚═╝ ╚═╝╚══════╝╚══════╝ ╚═╝ ╚═╝ ╚═════╝ ╚══════╝╚═╝ ╚═══╝ ╚═╝[/]"""
|
||
|
||
# ASCII Art - Hermes Caduceus (compact, fits in left panel)
|
||
HERMES_CADUCEUS = """[#CD7F32]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⢀⣀⡀⠀⣀⣀⠀⢀⣀⡀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#CD7F32]⠀⠀⠀⠀⠀⠀⢀⣠⣴⣾⣿⣿⣇⠸⣿⣿⠇⣸⣿⣿⣷⣦⣄⡀⠀⠀⠀⠀⠀⠀[/]
|
||
[#FFBF00]⠀⢀⣠⣴⣶⠿⠋⣩⡿⣿⡿⠻⣿⡇⢠⡄⢸⣿⠟⢿⣿⢿⣍⠙⠿⣶⣦⣄⡀⠀[/]
|
||
[#FFBF00]⠀⠀⠉⠉⠁⠶⠟⠋⠀⠉⠀⢀⣈⣁⡈⢁⣈⣁⡀⠀⠉⠀⠙⠻⠶⠈⠉⠉⠀⠀[/]
|
||
[#FFD700]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⣴⣿⡿⠛⢁⡈⠛⢿⣿⣦⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#FFD700]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠿⣿⣦⣤⣈⠁⢠⣴⣿⠿⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#FFBF00]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠈⠉⠻⢿⣿⣦⡉⠁⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#FFBF00]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠘⢷⣦⣈⠛⠃⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#CD7F32]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⢠⣴⠦⠈⠙⠿⣦⡄⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#CD7F32]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠸⣿⣤⡈⠁⢤⣿⠇⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#B8860B]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠉⠛⠷⠄⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#B8860B]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⢀⣀⠑⢶⣄⡀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#B8860B]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⣿⠁⢰⡆⠈⡿⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#B8860B]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠈⠳⠈⣡⠞⠁⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]
|
||
[#B8860B]⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠈⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀⠀[/]"""
|
||
|
||
# Compact banner for smaller terminals (fallback)
|
||
# Note: built dynamically by _build_compact_banner() to fit terminal width
|
||
COMPACT_BANNER = """
|
||
[bold #FFD700]╔══════════════════════════════════════════════════════════════╗[/]
|
||
[bold #FFD700]║[/] [#FFBF00]⚕ NOUS HERMES[/] [dim #B8860B]- AI Agent Framework[/] [bold #FFD700]║[/]
|
||
[bold #FFD700]║[/] [#CD7F32]Messenger of the Digital Gods[/] [dim #B8860B]Nous Research[/] [bold #FFD700]║[/]
|
||
[bold #FFD700]╚══════════════════════════════════════════════════════════════╝[/]
|
||
"""
|
||
|
||
|
||
def _build_compact_banner() -> str:
|
||
"""Build a compact banner that fits the current terminal width."""
|
||
w = min(shutil.get_terminal_size().columns - 2, 64)
|
||
if w < 30:
|
||
return "\n[#FFBF00]⚕ NOUS HERMES[/] [dim #B8860B]- Nous Research[/]\n"
|
||
inner = w - 2 # inside the box border
|
||
bar = "═" * w
|
||
line1 = "⚕ NOUS HERMES - AI Agent Framework"
|
||
line2 = "Messenger of the Digital Gods · Nous Research"
|
||
# Truncate and pad to fit
|
||
line1 = line1[:inner - 2].ljust(inner - 2)
|
||
line2 = line2[:inner - 2].ljust(inner - 2)
|
||
return (
|
||
f"\n[bold #FFD700]╔{bar}╗[/]\n"
|
||
f"[bold #FFD700]║[/] [#FFBF00]{line1}[/] [bold #FFD700]║[/]\n"
|
||
f"[bold #FFD700]║[/] [dim #B8860B]{line2}[/] [bold #FFD700]║[/]\n"
|
||
f"[bold #FFD700]╚{bar}╝[/]\n"
|
||
)
|
||
|
||
|
||
|
||
# ============================================================================
|
||
# Skill Slash Commands — dynamic commands generated from installed skills
|
||
# ============================================================================
|
||
|
||
from agent.skill_commands import (
|
||
scan_skill_commands,
|
||
get_skill_commands,
|
||
build_skill_invocation_message,
|
||
build_plan_path,
|
||
build_preloaded_skills_prompt,
|
||
)
|
||
|
||
_skill_commands = scan_skill_commands()
|
||
|
||
|
||
def _parse_skills_argument(skills: str | list[str] | tuple[str, ...] | None) -> list[str]:
|
||
"""Normalize a CLI skills flag into a deduplicated list of skill identifiers."""
|
||
if not skills:
|
||
return []
|
||
|
||
if isinstance(skills, str):
|
||
raw_values = [skills]
|
||
elif isinstance(skills, (list, tuple)):
|
||
raw_values = [str(item) for item in skills if item is not None]
|
||
else:
|
||
raw_values = [str(skills)]
|
||
|
||
parsed: list[str] = []
|
||
seen: set[str] = set()
|
||
for raw in raw_values:
|
||
for part in raw.split(","):
|
||
normalized = part.strip()
|
||
if not normalized or normalized in seen:
|
||
continue
|
||
seen.add(normalized)
|
||
parsed.append(normalized)
|
||
return parsed
|
||
|
||
|
||
def save_config_value(key_path: str, value: any) -> bool:
|
||
"""
|
||
Save a value to the active config file at the specified key path.
|
||
|
||
Respects the same lookup order as load_cli_config():
|
||
1. ~/.hermes/config.yaml (user config - preferred, used if it exists)
|
||
2. ./cli-config.yaml (project config - fallback)
|
||
|
||
Args:
|
||
key_path: Dot-separated path like "agent.system_prompt"
|
||
value: Value to save
|
||
|
||
Returns:
|
||
True if successful, False otherwise
|
||
"""
|
||
# Use the same precedence as load_cli_config: user config first, then project config
|
||
user_config_path = _hermes_home / 'config.yaml'
|
||
project_config_path = Path(__file__).parent / 'cli-config.yaml'
|
||
config_path = user_config_path if user_config_path.exists() else project_config_path
|
||
|
||
try:
|
||
# Ensure parent directory exists (for ~/.hermes/config.yaml on first use)
|
||
config_path.parent.mkdir(parents=True, exist_ok=True)
|
||
|
||
# Load existing config
|
||
if config_path.exists():
|
||
with open(config_path, 'r') as f:
|
||
config = yaml.safe_load(f) or {}
|
||
else:
|
||
config = {}
|
||
|
||
# Navigate to the key and set value
|
||
keys = key_path.split('.')
|
||
current = config
|
||
for key in keys[:-1]:
|
||
if key not in current or not isinstance(current[key], dict):
|
||
current[key] = {}
|
||
current = current[key]
|
||
current[keys[-1]] = value
|
||
|
||
# Save back
|
||
with open(config_path, 'w') as f:
|
||
yaml.dump(config, f, default_flow_style=False, sort_keys=False)
|
||
|
||
# Enforce owner-only permissions on config files (contain API keys)
|
||
try:
|
||
os.chmod(config_path, 0o600)
|
||
except (OSError, NotImplementedError):
|
||
pass
|
||
|
||
return True
|
||
except Exception as e:
|
||
logger.error("Failed to save config: %s", e)
|
||
return False
|
||
|
||
|
||
|
||
|
||
# ============================================================================
|
||
# HermesCLI Class
|
||
# ============================================================================
|
||
|
||
class HermesCLI:
|
||
"""
|
||
Interactive CLI for the Hermes Agent.
|
||
|
||
Provides a REPL interface with rich formatting, command history,
|
||
and tool execution capabilities.
|
||
"""
|
||
|
||
def __init__(
|
||
self,
|
||
model: str = None,
|
||
toolsets: List[str] = None,
|
||
provider: str = None,
|
||
api_key: str = None,
|
||
base_url: str = None,
|
||
max_turns: int = None,
|
||
verbose: bool = False,
|
||
compact: bool = False,
|
||
resume: str = None,
|
||
checkpoints: bool = False,
|
||
pass_session_id: bool = False,
|
||
):
|
||
"""
|
||
Initialize the Hermes CLI.
|
||
|
||
Args:
|
||
model: Model to use (default: from env or claude-sonnet)
|
||
toolsets: List of toolsets to enable (default: all)
|
||
provider: Inference provider ("auto", "openrouter", "nous", "openai-codex", "zai", "kimi-coding", "minimax", "minimax-cn")
|
||
api_key: API key (default: from environment)
|
||
base_url: API base URL (default: OpenRouter)
|
||
max_turns: Maximum tool-calling iterations shared with subagents (default: 90)
|
||
verbose: Enable verbose logging
|
||
compact: Use compact display mode
|
||
resume: Session ID to resume (restores conversation history from SQLite)
|
||
pass_session_id: Include the session ID in the agent's system prompt
|
||
"""
|
||
# Initialize Rich console
|
||
self.console = Console()
|
||
self.config = CLI_CONFIG
|
||
self.compact = compact if compact is not None else CLI_CONFIG["display"].get("compact", False)
|
||
# tool_progress: "off", "new", "all", "verbose" (from config.yaml display section)
|
||
self.tool_progress_mode = CLI_CONFIG["display"].get("tool_progress", "all")
|
||
# resume_display: "full" (show history) | "minimal" (one-liner only)
|
||
self.resume_display = CLI_CONFIG["display"].get("resume_display", "full")
|
||
# bell_on_complete: play terminal bell (\a) when agent finishes a response
|
||
self.bell_on_complete = CLI_CONFIG["display"].get("bell_on_complete", False)
|
||
# show_reasoning: display model thinking/reasoning before the response
|
||
self.show_reasoning = CLI_CONFIG["display"].get("show_reasoning", False)
|
||
|
||
self.verbose = verbose if verbose is not None else (self.tool_progress_mode == "verbose")
|
||
|
||
# streaming: stream tokens to the terminal as they arrive (display.streaming in config.yaml)
|
||
self.streaming_enabled = CLI_CONFIG["display"].get("streaming", False)
|
||
|
||
# Streaming display state
|
||
self._stream_buf = "" # Partial line buffer for line-buffered rendering
|
||
self._stream_started = False # True once first delta arrives
|
||
self._stream_box_opened = False # True once the response box header is printed
|
||
|
||
# Configuration - priority: CLI args > env vars > config file
|
||
# Model comes from: CLI arg or config.yaml (single source of truth).
|
||
# LLM_MODEL/OPENAI_MODEL env vars are NOT checked — config.yaml is
|
||
# authoritative. This avoids conflicts in multi-agent setups where
|
||
# env vars would stomp each other.
|
||
_model_config = CLI_CONFIG.get("model", {})
|
||
_config_model = _model_config.get("default", "") if isinstance(_model_config, dict) else (_model_config or "")
|
||
_FALLBACK_MODEL = "anthropic/claude-opus-4.6"
|
||
self.model = model or _config_model or _FALLBACK_MODEL
|
||
# Auto-detect model from local server if still on fallback
|
||
if self.model == _FALLBACK_MODEL:
|
||
_base_url = _model_config.get("base_url", "") if isinstance(_model_config, dict) else ""
|
||
if "localhost" in _base_url or "127.0.0.1" in _base_url:
|
||
from hermes_cli.runtime_provider import _auto_detect_local_model
|
||
_detected = _auto_detect_local_model(_base_url)
|
||
if _detected:
|
||
self.model = _detected
|
||
# Track whether model was explicitly chosen by the user or fell back
|
||
# to the global default. Provider-specific normalisation may override
|
||
# the default silently but should warn when overriding an explicit choice.
|
||
# A config model that matches the global fallback is NOT considered an
|
||
# explicit choice — the user just never changed it. But a config model
|
||
# like "gpt-5.3-codex" IS explicit and must be preserved.
|
||
self._model_is_default = not model and (
|
||
not _config_model or _config_model == _FALLBACK_MODEL
|
||
)
|
||
|
||
self._explicit_api_key = api_key
|
||
self._explicit_base_url = base_url
|
||
|
||
# Provider selection is resolved lazily at use-time via _ensure_runtime_credentials().
|
||
self.requested_provider = (
|
||
provider
|
||
or CLI_CONFIG["model"].get("provider")
|
||
or os.getenv("HERMES_INFERENCE_PROVIDER")
|
||
or "auto"
|
||
)
|
||
self._provider_source: Optional[str] = None
|
||
self.provider = self.requested_provider
|
||
self.api_mode = "chat_completions"
|
||
self.acp_command: Optional[str] = None
|
||
self.acp_args: list[str] = []
|
||
self.base_url = (
|
||
base_url
|
||
or os.getenv("OPENAI_BASE_URL")
|
||
or os.getenv("OPENROUTER_BASE_URL", CLI_CONFIG["model"]["base_url"])
|
||
)
|
||
# Match key to resolved base_url: OpenRouter URL → prefer OPENROUTER_API_KEY,
|
||
# custom endpoint → prefer OPENAI_API_KEY (issue #560).
|
||
# Note: _ensure_runtime_credentials() re-resolves this before first use.
|
||
if "openrouter.ai" in self.base_url:
|
||
self.api_key = api_key or os.getenv("OPENROUTER_API_KEY") or os.getenv("OPENAI_API_KEY")
|
||
else:
|
||
self.api_key = api_key or os.getenv("OPENAI_API_KEY") or os.getenv("OPENROUTER_API_KEY")
|
||
self._nous_key_expires_at: Optional[str] = None
|
||
self._nous_key_source: Optional[str] = None
|
||
# Max turns priority: CLI arg > config file > env var > default
|
||
if max_turns is not None: # CLI arg was explicitly set
|
||
self.max_turns = max_turns
|
||
elif CLI_CONFIG["agent"].get("max_turns"):
|
||
self.max_turns = CLI_CONFIG["agent"]["max_turns"]
|
||
elif CLI_CONFIG.get("max_turns"): # Backwards compat: root-level max_turns
|
||
self.max_turns = CLI_CONFIG["max_turns"]
|
||
elif os.getenv("HERMES_MAX_ITERATIONS"):
|
||
self.max_turns = int(os.getenv("HERMES_MAX_ITERATIONS"))
|
||
else:
|
||
self.max_turns = 90
|
||
|
||
# Parse and validate toolsets
|
||
self.enabled_toolsets = toolsets
|
||
if toolsets and "all" not in toolsets and "*" not in toolsets:
|
||
# Validate each toolset
|
||
invalid = [t for t in toolsets if not validate_toolset(t)]
|
||
if invalid:
|
||
self.console.print(f"[bold red]Warning: Unknown toolsets: {', '.join(invalid)}[/]")
|
||
|
||
# Filesystem checkpoints: CLI flag > config
|
||
cp_cfg = CLI_CONFIG.get("checkpoints", {})
|
||
if isinstance(cp_cfg, bool):
|
||
cp_cfg = {"enabled": cp_cfg}
|
||
self.checkpoints_enabled = checkpoints or cp_cfg.get("enabled", False)
|
||
self.checkpoint_max_snapshots = cp_cfg.get("max_snapshots", 50)
|
||
self.pass_session_id = pass_session_id
|
||
|
||
# Ephemeral system prompt: env var takes precedence, then config
|
||
self.system_prompt = (
|
||
os.getenv("HERMES_EPHEMERAL_SYSTEM_PROMPT", "")
|
||
or CLI_CONFIG["agent"].get("system_prompt", "")
|
||
)
|
||
self.personalities = CLI_CONFIG["agent"].get("personalities", {})
|
||
|
||
# Ephemeral prefill messages (few-shot priming, never persisted)
|
||
self.prefill_messages = _load_prefill_messages(
|
||
CLI_CONFIG["agent"].get("prefill_messages_file", "")
|
||
)
|
||
|
||
# Reasoning config (OpenRouter reasoning effort level)
|
||
self.reasoning_config = _parse_reasoning_config(
|
||
CLI_CONFIG["agent"].get("reasoning_effort", "")
|
||
)
|
||
|
||
# OpenRouter provider routing preferences
|
||
pr = CLI_CONFIG.get("provider_routing", {}) or {}
|
||
self._provider_sort = pr.get("sort")
|
||
self._providers_only = pr.get("only")
|
||
self._providers_ignore = pr.get("ignore")
|
||
self._providers_order = pr.get("order")
|
||
self._provider_require_params = pr.get("require_parameters", False)
|
||
self._provider_data_collection = pr.get("data_collection")
|
||
|
||
# Fallback model config — tried when primary provider fails after retries
|
||
fb = CLI_CONFIG.get("fallback_model") or {}
|
||
self._fallback_model = fb if fb.get("provider") and fb.get("model") else None
|
||
|
||
# Optional cheap-vs-strong routing for simple turns
|
||
self._smart_model_routing = CLI_CONFIG.get("smart_model_routing", {}) or {}
|
||
self._active_agent_route_signature = None
|
||
|
||
# Agent will be initialized on first use
|
||
self.agent: Optional[AIAgent] = None
|
||
self._app = None # prompt_toolkit Application (set in run())
|
||
|
||
# Conversation state
|
||
self.conversation_history: List[Dict[str, Any]] = []
|
||
self.session_start = datetime.now()
|
||
self._resumed = False
|
||
# Initialize SQLite session store early so /title works before first message
|
||
self._session_db = None
|
||
try:
|
||
from hermes_state import SessionDB
|
||
self._session_db = SessionDB()
|
||
except Exception:
|
||
pass
|
||
|
||
# Deferred title: stored in memory until the session is created in the DB
|
||
self._pending_title: Optional[str] = None
|
||
|
||
# Session ID: reuse existing one when resuming, otherwise generate fresh
|
||
if resume:
|
||
self.session_id = resume
|
||
self._resumed = True
|
||
else:
|
||
timestamp_str = self.session_start.strftime("%Y%m%d_%H%M%S")
|
||
short_uuid = uuid.uuid4().hex[:6]
|
||
self.session_id = f"{timestamp_str}_{short_uuid}"
|
||
|
||
# History file for persistent input recall across sessions
|
||
self._history_file = _hermes_home / ".hermes_history"
|
||
self._last_invalidate: float = 0.0 # throttle UI repaints
|
||
self._app = None
|
||
|
||
# State shared by interactive run() and single-query chat mode.
|
||
# These must exist before any direct chat() call because single-query
|
||
# mode does not go through run().
|
||
self._agent_running = False
|
||
self._pending_input = queue.Queue()
|
||
self._interrupt_queue = queue.Queue()
|
||
self._should_exit = False
|
||
self._last_ctrl_c_time = 0
|
||
self._clarify_state = None
|
||
self._clarify_freetext = False
|
||
self._clarify_deadline = 0
|
||
self._sudo_state = None
|
||
self._sudo_deadline = 0
|
||
self._approval_state = None
|
||
self._approval_deadline = 0
|
||
self._approval_lock = threading.Lock()
|
||
self._secret_state = None
|
||
self._secret_deadline = 0
|
||
self._spinner_text: str = "" # thinking spinner text for TUI
|
||
self._command_running = False
|
||
self._command_status = ""
|
||
self._attached_images: list[Path] = []
|
||
self._image_counter = 0
|
||
self.preloaded_skills: list[str] = []
|
||
self._startup_skills_line_shown = False
|
||
|
||
# Voice mode state (also reinitialized inside run() for interactive TUI).
|
||
self._voice_lock = threading.Lock()
|
||
self._voice_mode = False
|
||
self._voice_tts = False
|
||
self._voice_recorder = None
|
||
self._voice_recording = False
|
||
self._voice_processing = False
|
||
self._voice_continuous = False
|
||
self._voice_tts_done = threading.Event()
|
||
self._voice_tts_done.set()
|
||
|
||
# Status bar visibility (toggled via /statusbar)
|
||
self._status_bar_visible = True
|
||
|
||
# Background task tracking: {task_id: threading.Thread}
|
||
self._background_tasks: Dict[str, threading.Thread] = {}
|
||
self._background_task_counter = 0
|
||
|
||
def _invalidate(self, min_interval: float = 0.25) -> None:
|
||
"""Throttled UI repaint — prevents terminal blinking on slow/SSH connections."""
|
||
import time as _time
|
||
now = _time.monotonic()
|
||
if hasattr(self, "_app") and self._app and (now - self._last_invalidate) >= min_interval:
|
||
self._last_invalidate = now
|
||
self._app.invalidate()
|
||
|
||
def _status_bar_context_style(self, percent_used: Optional[int]) -> str:
|
||
if percent_used is None:
|
||
return "class:status-bar-dim"
|
||
if percent_used >= 95:
|
||
return "class:status-bar-critical"
|
||
if percent_used > 80:
|
||
return "class:status-bar-bad"
|
||
if percent_used >= 50:
|
||
return "class:status-bar-warn"
|
||
return "class:status-bar-good"
|
||
|
||
def _build_context_bar(self, percent_used: Optional[int], width: int = 10) -> str:
|
||
safe_percent = max(0, min(100, percent_used or 0))
|
||
filled = round((safe_percent / 100) * width)
|
||
return f"[{('█' * filled) + ('░' * max(0, width - filled))}]"
|
||
|
||
def _get_status_bar_snapshot(self) -> Dict[str, Any]:
|
||
model_name = self.model or "unknown"
|
||
model_short = model_name.split("/")[-1] if "/" in model_name else model_name
|
||
if model_short.endswith(".gguf"):
|
||
model_short = model_short[:-5]
|
||
if len(model_short) > 26:
|
||
model_short = f"{model_short[:23]}..."
|
||
|
||
elapsed_seconds = max(0.0, (datetime.now() - self.session_start).total_seconds())
|
||
snapshot = {
|
||
"model_name": model_name,
|
||
"model_short": model_short,
|
||
"duration": format_duration_compact(elapsed_seconds),
|
||
"context_tokens": 0,
|
||
"context_length": None,
|
||
"context_percent": None,
|
||
"session_input_tokens": 0,
|
||
"session_output_tokens": 0,
|
||
"session_cache_read_tokens": 0,
|
||
"session_cache_write_tokens": 0,
|
||
"session_prompt_tokens": 0,
|
||
"session_completion_tokens": 0,
|
||
"session_total_tokens": 0,
|
||
"session_api_calls": 0,
|
||
"compressions": 0,
|
||
}
|
||
|
||
agent = getattr(self, "agent", None)
|
||
if not agent:
|
||
return snapshot
|
||
|
||
snapshot["session_input_tokens"] = getattr(agent, "session_input_tokens", 0) or 0
|
||
snapshot["session_output_tokens"] = getattr(agent, "session_output_tokens", 0) or 0
|
||
snapshot["session_cache_read_tokens"] = getattr(agent, "session_cache_read_tokens", 0) or 0
|
||
snapshot["session_cache_write_tokens"] = getattr(agent, "session_cache_write_tokens", 0) or 0
|
||
snapshot["session_prompt_tokens"] = getattr(agent, "session_prompt_tokens", 0) or 0
|
||
snapshot["session_completion_tokens"] = getattr(agent, "session_completion_tokens", 0) or 0
|
||
snapshot["session_total_tokens"] = getattr(agent, "session_total_tokens", 0) or 0
|
||
snapshot["session_api_calls"] = getattr(agent, "session_api_calls", 0) or 0
|
||
|
||
compressor = getattr(agent, "context_compressor", None)
|
||
if compressor:
|
||
context_tokens = getattr(compressor, "last_prompt_tokens", 0) or 0
|
||
context_length = getattr(compressor, "context_length", 0) or 0
|
||
snapshot["context_tokens"] = context_tokens
|
||
snapshot["context_length"] = context_length or None
|
||
snapshot["compressions"] = getattr(compressor, "compression_count", 0) or 0
|
||
if context_length:
|
||
snapshot["context_percent"] = max(0, min(100, round((context_tokens / context_length) * 100)))
|
||
|
||
return snapshot
|
||
|
||
def _build_status_bar_text(self, width: Optional[int] = None) -> str:
|
||
try:
|
||
snapshot = self._get_status_bar_snapshot()
|
||
width = width or shutil.get_terminal_size((80, 24)).columns
|
||
percent = snapshot["context_percent"]
|
||
percent_label = f"{percent}%" if percent is not None else "--"
|
||
duration_label = snapshot["duration"]
|
||
|
||
if width < 52:
|
||
return f"⚕ {snapshot['model_short']} · {duration_label}"
|
||
if width < 76:
|
||
parts = [f"⚕ {snapshot['model_short']}", percent_label]
|
||
parts.append(duration_label)
|
||
return " · ".join(parts)
|
||
|
||
if snapshot["context_length"]:
|
||
ctx_total = _format_context_length(snapshot["context_length"])
|
||
ctx_used = format_token_count_compact(snapshot["context_tokens"])
|
||
context_label = f"{ctx_used}/{ctx_total}"
|
||
else:
|
||
context_label = "ctx --"
|
||
|
||
parts = [f"⚕ {snapshot['model_short']}", context_label, percent_label]
|
||
parts.append(duration_label)
|
||
return " │ ".join(parts)
|
||
except Exception:
|
||
return f"⚕ {self.model if getattr(self, 'model', None) else 'Hermes'}"
|
||
|
||
def _get_status_bar_fragments(self):
|
||
if not self._status_bar_visible:
|
||
return []
|
||
try:
|
||
snapshot = self._get_status_bar_snapshot()
|
||
width = shutil.get_terminal_size((80, 24)).columns
|
||
duration_label = snapshot["duration"]
|
||
|
||
if width < 52:
|
||
return [
|
||
("class:status-bar", " ⚕ "),
|
||
("class:status-bar-strong", snapshot["model_short"]),
|
||
("class:status-bar-dim", " · "),
|
||
("class:status-bar-dim", duration_label),
|
||
("class:status-bar", " "),
|
||
]
|
||
|
||
percent = snapshot["context_percent"]
|
||
percent_label = f"{percent}%" if percent is not None else "--"
|
||
if width < 76:
|
||
frags = [
|
||
("class:status-bar", " ⚕ "),
|
||
("class:status-bar-strong", snapshot["model_short"]),
|
||
("class:status-bar-dim", " · "),
|
||
(self._status_bar_context_style(percent), percent_label),
|
||
]
|
||
frags.extend([
|
||
("class:status-bar-dim", " · "),
|
||
("class:status-bar-dim", duration_label),
|
||
("class:status-bar", " "),
|
||
])
|
||
return frags
|
||
|
||
if snapshot["context_length"]:
|
||
ctx_total = _format_context_length(snapshot["context_length"])
|
||
ctx_used = format_token_count_compact(snapshot["context_tokens"])
|
||
context_label = f"{ctx_used}/{ctx_total}"
|
||
else:
|
||
context_label = "ctx --"
|
||
|
||
bar_style = self._status_bar_context_style(percent)
|
||
frags = [
|
||
("class:status-bar", " ⚕ "),
|
||
("class:status-bar-strong", snapshot["model_short"]),
|
||
("class:status-bar-dim", " │ "),
|
||
("class:status-bar-dim", context_label),
|
||
("class:status-bar-dim", " │ "),
|
||
(bar_style, self._build_context_bar(percent)),
|
||
("class:status-bar-dim", " "),
|
||
(bar_style, percent_label),
|
||
]
|
||
frags.extend([
|
||
("class:status-bar-dim", " │ "),
|
||
("class:status-bar-dim", duration_label),
|
||
("class:status-bar", " "),
|
||
])
|
||
return frags
|
||
except Exception:
|
||
return [("class:status-bar", f" {self._build_status_bar_text()} ")]
|
||
|
||
def _normalize_model_for_provider(self, resolved_provider: str) -> bool:
|
||
"""Normalize provider-specific model IDs and routing."""
|
||
current_model = (self.model or "").strip()
|
||
changed = False
|
||
|
||
if resolved_provider == "copilot":
|
||
try:
|
||
from hermes_cli.models import copilot_model_api_mode, normalize_copilot_model_id
|
||
|
||
canonical = normalize_copilot_model_id(current_model, api_key=self.api_key)
|
||
if canonical and canonical != current_model:
|
||
if not self._model_is_default:
|
||
self.console.print(
|
||
f"[yellow]⚠️ Normalized Copilot model '{current_model}' to '{canonical}'.[/]"
|
||
)
|
||
self.model = canonical
|
||
current_model = canonical
|
||
changed = True
|
||
|
||
resolved_mode = copilot_model_api_mode(current_model, api_key=self.api_key)
|
||
if resolved_mode != self.api_mode:
|
||
self.api_mode = resolved_mode
|
||
changed = True
|
||
except Exception:
|
||
pass
|
||
return changed
|
||
|
||
if resolved_provider != "openai-codex":
|
||
return False
|
||
|
||
# 1. Strip provider prefix ("openai/gpt-5.4" → "gpt-5.4")
|
||
if "/" in current_model:
|
||
slug = current_model.split("/", 1)[1]
|
||
if not self._model_is_default:
|
||
self.console.print(
|
||
f"[yellow]⚠️ Stripped provider prefix from '{current_model}'; "
|
||
f"using '{slug}' for OpenAI Codex.[/]"
|
||
)
|
||
self.model = slug
|
||
current_model = slug
|
||
changed = True
|
||
|
||
# 2. Replace untouched default with a Codex model
|
||
if self._model_is_default:
|
||
fallback_model = "gpt-5.3-codex"
|
||
try:
|
||
from hermes_cli.codex_models import get_codex_model_ids
|
||
|
||
available = get_codex_model_ids(
|
||
access_token=self.api_key if self.api_key else None,
|
||
)
|
||
if available:
|
||
fallback_model = available[0]
|
||
except Exception:
|
||
pass
|
||
|
||
if current_model != fallback_model:
|
||
self.model = fallback_model
|
||
changed = True
|
||
|
||
return changed
|
||
|
||
def _on_thinking(self, text: str) -> None:
|
||
"""Called by agent when thinking starts/stops. Updates TUI spinner."""
|
||
self._spinner_text = text or ""
|
||
self._invalidate()
|
||
|
||
# ── Streaming display ────────────────────────────────────────────────
|
||
|
||
def _stream_reasoning_delta(self, text: str) -> None:
|
||
"""Stream reasoning/thinking tokens into a dim box above the response.
|
||
|
||
Opens a dim reasoning box on first token, streams line-by-line.
|
||
The box is closed automatically when content tokens start arriving
|
||
(via _stream_delta → _emit_stream_text).
|
||
"""
|
||
if not text:
|
||
return
|
||
|
||
# Open reasoning box on first reasoning token
|
||
if not getattr(self, "_reasoning_box_opened", False):
|
||
self._reasoning_box_opened = True
|
||
w = shutil.get_terminal_size().columns
|
||
r_label = " Reasoning "
|
||
r_fill = w - 2 - len(r_label)
|
||
_cprint(f"\n{_DIM}┌─{r_label}{'─' * max(r_fill - 1, 0)}┐{_RST}")
|
||
|
||
self._reasoning_buf = getattr(self, "_reasoning_buf", "") + text
|
||
|
||
# Emit complete lines
|
||
while "\n" in self._reasoning_buf:
|
||
line, self._reasoning_buf = self._reasoning_buf.split("\n", 1)
|
||
_cprint(f"{_DIM}{line}{_RST}")
|
||
|
||
def _close_reasoning_box(self) -> None:
|
||
"""Close the live reasoning box if it's open."""
|
||
if getattr(self, "_reasoning_box_opened", False):
|
||
# Flush remaining reasoning buffer
|
||
buf = getattr(self, "_reasoning_buf", "")
|
||
if buf:
|
||
_cprint(f"{_DIM}{buf}{_RST}")
|
||
self._reasoning_buf = ""
|
||
w = shutil.get_terminal_size().columns
|
||
_cprint(f"{_DIM}└{'─' * (w - 2)}┘{_RST}")
|
||
self._reasoning_box_opened = False
|
||
|
||
def _stream_delta(self, text) -> None:
|
||
"""Line-buffered streaming callback for real-time token rendering.
|
||
|
||
Receives text deltas from the agent as tokens arrive. Buffers
|
||
partial lines and emits complete lines via _cprint to work
|
||
reliably with prompt_toolkit's patch_stdout.
|
||
|
||
Reasoning/thinking blocks (<REASONING_SCRATCHPAD>, <think>, etc.)
|
||
are suppressed during streaming since they'd display raw XML tags.
|
||
The agent strips them from the final response anyway.
|
||
|
||
A ``None`` value signals an intermediate turn boundary (tools are
|
||
about to execute). Flushes any open boxes and resets state so
|
||
tool feed lines render cleanly between turns.
|
||
"""
|
||
if text is None:
|
||
self._flush_stream()
|
||
self._reset_stream_state()
|
||
return
|
||
if not text:
|
||
return
|
||
|
||
self._stream_started = True
|
||
|
||
# ── Tag-based reasoning suppression ──
|
||
# Track whether we're inside a reasoning/thinking block.
|
||
# These tags are model-generated (system prompt tells the model
|
||
# to use them) and get stripped from final_response. We must
|
||
# suppress them during streaming too — unless show_reasoning is
|
||
# enabled, in which case we route the inner content to the
|
||
# reasoning display box instead of discarding it.
|
||
_OPEN_TAGS = ("<REASONING_SCRATCHPAD>", "<think>", "<reasoning>", "<THINKING>", "<thinking>")
|
||
_CLOSE_TAGS = ("</REASONING_SCRATCHPAD>", "</think>", "</reasoning>", "</THINKING>", "</thinking>")
|
||
|
||
# Append to a pre-filter buffer first
|
||
self._stream_prefilt = getattr(self, "_stream_prefilt", "") + text
|
||
|
||
# Check if we're entering a reasoning block
|
||
if not getattr(self, "_in_reasoning_block", False):
|
||
for tag in _OPEN_TAGS:
|
||
idx = self._stream_prefilt.find(tag)
|
||
if idx != -1:
|
||
# Emit everything before the tag
|
||
before = self._stream_prefilt[:idx]
|
||
if before:
|
||
self._emit_stream_text(before)
|
||
self._in_reasoning_block = True
|
||
self._stream_prefilt = self._stream_prefilt[idx + len(tag):]
|
||
break
|
||
|
||
# Could also be a partial open tag at the end — hold it back
|
||
if not getattr(self, "_in_reasoning_block", False):
|
||
# Check for partial tag match at the end
|
||
safe = self._stream_prefilt
|
||
for tag in _OPEN_TAGS:
|
||
for i in range(1, len(tag)):
|
||
if self._stream_prefilt.endswith(tag[:i]):
|
||
safe = self._stream_prefilt[:-i]
|
||
break
|
||
if safe:
|
||
self._emit_stream_text(safe)
|
||
self._stream_prefilt = self._stream_prefilt[len(safe):]
|
||
return
|
||
|
||
# Inside a reasoning block — look for close tag.
|
||
# Keep accumulating _stream_prefilt because close tags can arrive
|
||
# split across multiple tokens (e.g. "</REASONING_SCRATCH" + "PAD>...").
|
||
if getattr(self, "_in_reasoning_block", False):
|
||
for tag in _CLOSE_TAGS:
|
||
idx = self._stream_prefilt.find(tag)
|
||
if idx != -1:
|
||
self._in_reasoning_block = False
|
||
# When show_reasoning is on, route inner content to
|
||
# the reasoning display box instead of discarding.
|
||
if self.show_reasoning:
|
||
inner = self._stream_prefilt[:idx]
|
||
if inner:
|
||
self._stream_reasoning_delta(inner)
|
||
after = self._stream_prefilt[idx + len(tag):]
|
||
self._stream_prefilt = ""
|
||
# Process remaining text after close tag through full
|
||
# filtering (it could contain another open tag)
|
||
if after:
|
||
self._stream_delta(after)
|
||
return
|
||
# When show_reasoning is on, stream reasoning content live
|
||
# instead of silently accumulating. Keep only the tail that
|
||
# could be a partial close tag prefix.
|
||
max_tag_len = max(len(t) for t in _CLOSE_TAGS)
|
||
if len(self._stream_prefilt) > max_tag_len:
|
||
if self.show_reasoning:
|
||
# Route the safe prefix to reasoning display
|
||
safe_reasoning = self._stream_prefilt[:-max_tag_len]
|
||
self._stream_reasoning_delta(safe_reasoning)
|
||
self._stream_prefilt = self._stream_prefilt[-max_tag_len:]
|
||
return
|
||
|
||
def _emit_stream_text(self, text: str) -> None:
|
||
"""Emit filtered text to the streaming display."""
|
||
if not text:
|
||
return
|
||
|
||
# Close the live reasoning box before opening the response box
|
||
self._close_reasoning_box()
|
||
|
||
# Open the response box header on the very first visible text
|
||
if not self._stream_box_opened:
|
||
# Strip leading whitespace/newlines before first visible content
|
||
text = text.lstrip("\n")
|
||
if not text:
|
||
return
|
||
self._stream_box_opened = True
|
||
try:
|
||
from hermes_cli.skin_engine import get_active_skin
|
||
_skin = get_active_skin()
|
||
label = _skin.get_branding("response_label", "⚕ Hermes")
|
||
except Exception:
|
||
label = "⚕ Hermes"
|
||
w = shutil.get_terminal_size().columns
|
||
fill = w - 2 - len(label)
|
||
_cprint(f"\n{_GOLD}╭─{label}{'─' * max(fill - 1, 0)}╮{_RST}")
|
||
|
||
self._stream_buf += text
|
||
|
||
# Emit complete lines, keep partial remainder in buffer
|
||
while "\n" in self._stream_buf:
|
||
line, self._stream_buf = self._stream_buf.split("\n", 1)
|
||
_cprint(line)
|
||
|
||
def _flush_stream(self) -> None:
|
||
"""Emit any remaining partial line from the stream buffer and close the box."""
|
||
# Close reasoning box if still open (in case no content tokens arrived)
|
||
self._close_reasoning_box()
|
||
|
||
if self._stream_buf:
|
||
_cprint(self._stream_buf)
|
||
self._stream_buf = ""
|
||
|
||
# Close the response box
|
||
if self._stream_box_opened:
|
||
w = shutil.get_terminal_size().columns
|
||
_cprint(f"{_GOLD}╰{'─' * (w - 2)}╯{_RST}")
|
||
|
||
def _reset_stream_state(self) -> None:
|
||
"""Reset streaming state before each agent invocation."""
|
||
self._stream_buf = ""
|
||
self._stream_started = False
|
||
self._stream_box_opened = False
|
||
self._stream_prefilt = ""
|
||
self._in_reasoning_block = False
|
||
self._reasoning_box_opened = False
|
||
self._reasoning_buf = ""
|
||
|
||
def _slow_command_status(self, command: str) -> str:
|
||
"""Return a user-facing status message for slower slash commands."""
|
||
cmd_lower = command.lower().strip()
|
||
if cmd_lower.startswith("/skills search"):
|
||
return "Searching skills..."
|
||
if cmd_lower.startswith("/skills browse"):
|
||
return "Loading skills..."
|
||
if cmd_lower.startswith("/skills inspect"):
|
||
return "Inspecting skill..."
|
||
if cmd_lower.startswith("/skills install"):
|
||
return "Installing skill..."
|
||
if cmd_lower.startswith("/skills"):
|
||
return "Processing skills command..."
|
||
if cmd_lower == "/reload-mcp":
|
||
return "Reloading MCP servers..."
|
||
if cmd_lower.startswith("/browser"):
|
||
return "Configuring browser..."
|
||
return "Processing command..."
|
||
|
||
def _command_spinner_frame(self) -> str:
|
||
"""Return the current spinner frame for slow slash commands."""
|
||
import time as _time
|
||
|
||
frame_idx = int(_time.monotonic() * 10) % len(_COMMAND_SPINNER_FRAMES)
|
||
return _COMMAND_SPINNER_FRAMES[frame_idx]
|
||
|
||
@contextmanager
|
||
def _busy_command(self, status: str):
|
||
"""Expose a temporary busy state in the TUI while a slash command runs."""
|
||
self._command_running = True
|
||
self._command_status = status
|
||
self._invalidate(min_interval=0.0)
|
||
try:
|
||
print(f"⏳ {status}")
|
||
yield
|
||
finally:
|
||
self._command_running = False
|
||
self._command_status = ""
|
||
self._invalidate(min_interval=0.0)
|
||
|
||
def _ensure_runtime_credentials(self) -> bool:
|
||
"""
|
||
Ensure runtime credentials are resolved before agent use.
|
||
Re-resolves provider credentials so key rotation and token refresh
|
||
are picked up without restarting the CLI.
|
||
Returns True if credentials are ready, False on auth failure.
|
||
"""
|
||
from hermes_cli.runtime_provider import (
|
||
resolve_runtime_provider,
|
||
format_runtime_provider_error,
|
||
)
|
||
|
||
try:
|
||
runtime = resolve_runtime_provider(
|
||
requested=self.requested_provider,
|
||
explicit_api_key=self._explicit_api_key,
|
||
explicit_base_url=self._explicit_base_url,
|
||
)
|
||
except Exception as exc:
|
||
message = format_runtime_provider_error(exc)
|
||
self.console.print(f"[bold red]{message}[/]")
|
||
return False
|
||
|
||
api_key = runtime.get("api_key")
|
||
base_url = runtime.get("base_url")
|
||
resolved_provider = runtime.get("provider", "openrouter")
|
||
resolved_api_mode = runtime.get("api_mode", self.api_mode)
|
||
resolved_acp_command = runtime.get("command")
|
||
resolved_acp_args = list(runtime.get("args") or [])
|
||
if not isinstance(api_key, str) or not api_key:
|
||
self.console.print("[bold red]Provider resolver returned an empty API key.[/]")
|
||
return False
|
||
if not isinstance(base_url, str) or not base_url:
|
||
self.console.print("[bold red]Provider resolver returned an empty base URL.[/]")
|
||
return False
|
||
|
||
credentials_changed = api_key != self.api_key or base_url != self.base_url
|
||
routing_changed = (
|
||
resolved_provider != self.provider
|
||
or resolved_api_mode != self.api_mode
|
||
or resolved_acp_command != self.acp_command
|
||
or resolved_acp_args != self.acp_args
|
||
)
|
||
self.provider = resolved_provider
|
||
self.api_mode = resolved_api_mode
|
||
self.acp_command = resolved_acp_command
|
||
self.acp_args = resolved_acp_args
|
||
self._provider_source = runtime.get("source")
|
||
self.api_key = api_key
|
||
self.base_url = base_url
|
||
|
||
# Normalize model for the resolved provider (e.g. swap non-Codex
|
||
# models when provider is openai-codex). Fixes #651.
|
||
model_changed = self._normalize_model_for_provider(resolved_provider)
|
||
|
||
# AIAgent/OpenAI client holds auth at init time, so rebuild if key,
|
||
# routing, or the effective model changed.
|
||
if (credentials_changed or routing_changed or model_changed) and self.agent is not None:
|
||
self.agent = None
|
||
self._active_agent_route_signature = None
|
||
|
||
return True
|
||
|
||
def _resolve_turn_agent_config(self, user_message: str) -> dict:
|
||
"""Resolve model/runtime overrides for a single user turn."""
|
||
from agent.smart_model_routing import resolve_turn_route
|
||
|
||
return resolve_turn_route(
|
||
user_message,
|
||
self._smart_model_routing,
|
||
{
|
||
"model": self.model,
|
||
"api_key": self.api_key,
|
||
"base_url": self.base_url,
|
||
"provider": self.provider,
|
||
"api_mode": self.api_mode,
|
||
"command": self.acp_command,
|
||
"args": list(self.acp_args or []),
|
||
},
|
||
)
|
||
|
||
def _init_agent(self, *, model_override: str = None, runtime_override: dict = None, route_label: str = None) -> bool:
|
||
"""
|
||
Initialize the agent on first use.
|
||
When resuming a session, restores conversation history from SQLite.
|
||
|
||
Returns:
|
||
bool: True if successful, False otherwise
|
||
"""
|
||
if self.agent is not None:
|
||
return True
|
||
|
||
if not self._ensure_runtime_credentials():
|
||
return False
|
||
|
||
# Initialize SQLite session store for CLI sessions (if not already done in __init__)
|
||
if self._session_db is None:
|
||
try:
|
||
from hermes_state import SessionDB
|
||
self._session_db = SessionDB()
|
||
except Exception as e:
|
||
logger.debug("SQLite session store not available: %s", e)
|
||
|
||
# If resuming, validate the session exists and load its history.
|
||
# _preload_resumed_session() may have already loaded it (called from
|
||
# run() for immediate display). In that case, conversation_history
|
||
# is non-empty and we skip the DB round-trip.
|
||
if self._resumed and self._session_db and not self.conversation_history:
|
||
session_meta = self._session_db.get_session(self.session_id)
|
||
if not session_meta:
|
||
_cprint(f"\033[1;31mSession not found: {self.session_id}{_RST}")
|
||
_cprint(f"{_DIM}Use a session ID from a previous CLI run (hermes sessions list).{_RST}")
|
||
return False
|
||
restored = self._session_db.get_messages_as_conversation(self.session_id)
|
||
if restored:
|
||
self.conversation_history = restored
|
||
msg_count = len([m for m in restored if m.get("role") == "user"])
|
||
title_part = ""
|
||
if session_meta.get("title"):
|
||
title_part = f" \"{session_meta['title']}\""
|
||
ChatConsole().print(
|
||
f"[bold {_accent_hex()}]↻ Resumed session[/] "
|
||
f"[bold]{_escape(self.session_id)}[/]"
|
||
f"[bold {_accent_hex()}]{_escape(title_part)}[/] "
|
||
f"({msg_count} user message{'s' if msg_count != 1 else ''}, {len(restored)} total messages)"
|
||
)
|
||
else:
|
||
ChatConsole().print(
|
||
f"[bold {_accent_hex()}]Session {_escape(self.session_id)} found but has no messages. Starting fresh.[/]"
|
||
)
|
||
# Re-open the session (clear ended_at so it's active again)
|
||
try:
|
||
self._session_db._conn.execute(
|
||
"UPDATE sessions SET ended_at = NULL, end_reason = NULL WHERE id = ?",
|
||
(self.session_id,),
|
||
)
|
||
self._session_db._conn.commit()
|
||
except Exception:
|
||
pass
|
||
|
||
try:
|
||
runtime = runtime_override or {
|
||
"api_key": self.api_key,
|
||
"base_url": self.base_url,
|
||
"provider": self.provider,
|
||
"api_mode": self.api_mode,
|
||
"command": self.acp_command,
|
||
"args": list(self.acp_args or []),
|
||
}
|
||
effective_model = model_override or self.model
|
||
self.agent = AIAgent(
|
||
model=effective_model,
|
||
api_key=runtime.get("api_key"),
|
||
base_url=runtime.get("base_url"),
|
||
provider=runtime.get("provider"),
|
||
api_mode=runtime.get("api_mode"),
|
||
acp_command=runtime.get("command"),
|
||
acp_args=runtime.get("args"),
|
||
max_iterations=self.max_turns,
|
||
enabled_toolsets=self.enabled_toolsets,
|
||
verbose_logging=self.verbose,
|
||
quiet_mode=not self.verbose,
|
||
ephemeral_system_prompt=self.system_prompt if self.system_prompt else None,
|
||
prefill_messages=self.prefill_messages or None,
|
||
reasoning_config=self.reasoning_config,
|
||
providers_allowed=self._providers_only,
|
||
providers_ignored=self._providers_ignore,
|
||
providers_order=self._providers_order,
|
||
provider_sort=self._provider_sort,
|
||
provider_require_parameters=self._provider_require_params,
|
||
provider_data_collection=self._provider_data_collection,
|
||
session_id=self.session_id,
|
||
platform="cli",
|
||
session_db=self._session_db,
|
||
clarify_callback=self._clarify_callback,
|
||
reasoning_callback=(
|
||
self._stream_reasoning_delta if (self.streaming_enabled and self.show_reasoning)
|
||
else self._on_reasoning if (self.show_reasoning or self.verbose)
|
||
else None
|
||
),
|
||
honcho_session_key=None, # resolved by run_agent via config sessions map / title
|
||
fallback_model=self._fallback_model,
|
||
thinking_callback=self._on_thinking,
|
||
checkpoints_enabled=self.checkpoints_enabled,
|
||
checkpoint_max_snapshots=self.checkpoint_max_snapshots,
|
||
pass_session_id=self.pass_session_id,
|
||
tool_progress_callback=self._on_tool_progress,
|
||
stream_delta_callback=self._stream_delta if self.streaming_enabled else None,
|
||
)
|
||
self._active_agent_route_signature = (
|
||
effective_model,
|
||
runtime.get("provider"),
|
||
runtime.get("base_url"),
|
||
runtime.get("api_mode"),
|
||
runtime.get("command"),
|
||
tuple(runtime.get("args") or ()),
|
||
)
|
||
|
||
if self._pending_title and self._session_db:
|
||
try:
|
||
self._session_db.set_session_title(self.session_id, self._pending_title)
|
||
_cprint(f" Session title applied: {self._pending_title}")
|
||
self._pending_title = None
|
||
except (ValueError, Exception) as e:
|
||
_cprint(f" Could not apply pending title: {e}")
|
||
self._pending_title = None
|
||
return True
|
||
except Exception as e:
|
||
self.console.print(f"[bold red]Failed to initialize agent: {e}[/]")
|
||
return False
|
||
|
||
def show_banner(self):
|
||
"""Display the welcome banner in Claude Code style."""
|
||
self.console.clear()
|
||
if self.preloaded_skills and not self._startup_skills_line_shown:
|
||
skills_label = ", ".join(self.preloaded_skills)
|
||
self.console.print(
|
||
f"[bold {_accent_hex()}]Activated skills:[/] {skills_label}"
|
||
)
|
||
self.console.print()
|
||
self._startup_skills_line_shown = True
|
||
|
||
# Auto-compact for narrow terminals — the full banner with caduceus
|
||
# + tool list needs ~80 columns minimum to render without wrapping.
|
||
term_width = shutil.get_terminal_size().columns
|
||
use_compact = self.compact or term_width < 80
|
||
|
||
if use_compact:
|
||
self.console.print(_build_compact_banner())
|
||
self._show_status()
|
||
else:
|
||
# Get tools for display
|
||
tools = get_tool_definitions(enabled_toolsets=self.enabled_toolsets, quiet_mode=True)
|
||
|
||
# Get terminal working directory (where commands will execute)
|
||
cwd = os.getenv("TERMINAL_CWD", os.getcwd())
|
||
|
||
# Get context length for display
|
||
ctx_len = None
|
||
if hasattr(self, 'agent') and self.agent and hasattr(self.agent, 'context_compressor'):
|
||
ctx_len = self.agent.context_compressor.context_length
|
||
|
||
# Build and display the banner
|
||
build_welcome_banner(
|
||
console=self.console,
|
||
model=self.model,
|
||
cwd=cwd,
|
||
tools=tools,
|
||
enabled_toolsets=self.enabled_toolsets,
|
||
session_id=self.session_id,
|
||
context_length=ctx_len,
|
||
)
|
||
|
||
# Show tool availability warnings if any tools are disabled
|
||
self._show_tool_availability_warnings()
|
||
|
||
self.console.print()
|
||
|
||
def _preload_resumed_session(self) -> bool:
|
||
"""Load a resumed session's history from the DB early (before first chat).
|
||
|
||
Called from run() so the conversation history is available for display
|
||
before the user sends their first message. Sets
|
||
``self.conversation_history`` and prints the one-liner status. Returns
|
||
True if history was loaded, False otherwise.
|
||
|
||
The corresponding block in ``_init_agent()`` checks whether history is
|
||
already populated and skips the DB round-trip.
|
||
"""
|
||
if not self._resumed or not self._session_db:
|
||
return False
|
||
|
||
session_meta = self._session_db.get_session(self.session_id)
|
||
if not session_meta:
|
||
self.console.print(
|
||
f"[bold red]Session not found: {self.session_id}[/]"
|
||
)
|
||
self.console.print(
|
||
"[dim]Use a session ID from a previous CLI run "
|
||
"(hermes sessions list).[/]"
|
||
)
|
||
return False
|
||
|
||
restored = self._session_db.get_messages_as_conversation(self.session_id)
|
||
if restored:
|
||
self.conversation_history = restored
|
||
msg_count = len([m for m in restored if m.get("role") == "user"])
|
||
title_part = ""
|
||
if session_meta.get("title"):
|
||
title_part = f' "{session_meta["title"]}"'
|
||
self.console.print(
|
||
f"[#DAA520]↻ Resumed session [bold]{self.session_id}[/bold]"
|
||
f"{title_part} "
|
||
f"({msg_count} user message{'s' if msg_count != 1 else ''}, "
|
||
f"{len(restored)} total messages)[/]"
|
||
)
|
||
else:
|
||
self.console.print(
|
||
f"[#DAA520]Session {self.session_id} found but has no "
|
||
f"messages. Starting fresh.[/]"
|
||
)
|
||
return False
|
||
|
||
# Re-open the session (clear ended_at so it's active again)
|
||
try:
|
||
self._session_db._conn.execute(
|
||
"UPDATE sessions SET ended_at = NULL, end_reason = NULL "
|
||
"WHERE id = ?",
|
||
(self.session_id,),
|
||
)
|
||
self._session_db._conn.commit()
|
||
except Exception:
|
||
pass
|
||
|
||
return True
|
||
|
||
def _display_resumed_history(self):
|
||
"""Render a compact recap of previous conversation messages.
|
||
|
||
Uses Rich markup with dim/muted styling so the recap is visually
|
||
distinct from the active conversation. Caps the display at the
|
||
last ``MAX_DISPLAY_EXCHANGES`` user/assistant exchanges and shows
|
||
an indicator for earlier hidden messages.
|
||
"""
|
||
if not self.conversation_history:
|
||
return
|
||
|
||
# Check config: resume_display setting
|
||
if self.resume_display == "minimal":
|
||
return
|
||
|
||
MAX_DISPLAY_EXCHANGES = 10 # max user+assistant pairs to show
|
||
MAX_USER_LEN = 300 # truncate user messages
|
||
MAX_ASST_LEN = 200 # truncate assistant text
|
||
MAX_ASST_LINES = 3 # max lines of assistant text
|
||
|
||
def _strip_reasoning(text: str) -> str:
|
||
"""Remove <REASONING_SCRATCHPAD>...</REASONING_SCRATCHPAD> blocks
|
||
from displayed text (reasoning model internal thoughts)."""
|
||
import re
|
||
cleaned = re.sub(
|
||
r"<REASONING_SCRATCHPAD>.*?</REASONING_SCRATCHPAD>\s*",
|
||
"", text, flags=re.DOTALL,
|
||
)
|
||
# Also strip unclosed reasoning tags at the end
|
||
cleaned = re.sub(
|
||
r"<REASONING_SCRATCHPAD>.*$",
|
||
"", cleaned, flags=re.DOTALL,
|
||
)
|
||
return cleaned.strip()
|
||
|
||
# Collect displayable entries (skip system, tool-result messages)
|
||
entries = [] # list of (role, display_text)
|
||
for msg in self.conversation_history:
|
||
role = msg.get("role", "")
|
||
content = msg.get("content")
|
||
tool_calls = msg.get("tool_calls") or []
|
||
|
||
if role == "system":
|
||
continue
|
||
if role == "tool":
|
||
continue
|
||
|
||
if role == "user":
|
||
text = "" if content is None else str(content)
|
||
# Handle multimodal content (list of dicts)
|
||
if isinstance(content, list):
|
||
parts = []
|
||
for part in content:
|
||
if isinstance(part, dict) and part.get("type") == "text":
|
||
parts.append(part.get("text", ""))
|
||
elif isinstance(part, dict) and part.get("type") == "image_url":
|
||
parts.append("[image]")
|
||
text = " ".join(parts)
|
||
if len(text) > MAX_USER_LEN:
|
||
text = text[:MAX_USER_LEN] + "..."
|
||
entries.append(("user", text))
|
||
|
||
elif role == "assistant":
|
||
text = "" if content is None else str(content)
|
||
text = _strip_reasoning(text)
|
||
parts = []
|
||
if text:
|
||
lines = text.splitlines()
|
||
if len(lines) > MAX_ASST_LINES:
|
||
text = "\n".join(lines[:MAX_ASST_LINES]) + " ..."
|
||
if len(text) > MAX_ASST_LEN:
|
||
text = text[:MAX_ASST_LEN] + "..."
|
||
parts.append(text)
|
||
if tool_calls:
|
||
tc_count = len(tool_calls)
|
||
# Extract tool names
|
||
names = []
|
||
for tc in tool_calls:
|
||
fn = tc.get("function", {})
|
||
name = fn.get("name", "unknown") if isinstance(fn, dict) else "unknown"
|
||
if name not in names:
|
||
names.append(name)
|
||
names_str = ", ".join(names[:4])
|
||
if len(names) > 4:
|
||
names_str += ", ..."
|
||
noun = "call" if tc_count == 1 else "calls"
|
||
parts.append(f"[{tc_count} tool {noun}: {names_str}]")
|
||
if not parts:
|
||
# Skip pure-reasoning messages that have no visible output
|
||
continue
|
||
entries.append(("assistant", " ".join(parts)))
|
||
|
||
if not entries:
|
||
return
|
||
|
||
# Determine if we need to truncate
|
||
skipped = 0
|
||
if len(entries) > MAX_DISPLAY_EXCHANGES * 2:
|
||
skipped = len(entries) - MAX_DISPLAY_EXCHANGES * 2
|
||
entries = entries[skipped:]
|
||
|
||
# Build the display using Rich
|
||
from rich.panel import Panel
|
||
from rich.text import Text
|
||
|
||
try:
|
||
from hermes_cli.skin_engine import get_active_skin
|
||
_skin = get_active_skin()
|
||
_history_text_c = _skin.get_color("banner_text", "#FFF8DC")
|
||
_session_label_c = _skin.get_color("session_label", "#DAA520")
|
||
_session_border_c = _skin.get_color("session_border", "#8B8682")
|
||
_assistant_label_c = _skin.get_color("ui_ok", "#8FBC8F")
|
||
except Exception:
|
||
_history_text_c = "#FFF8DC"
|
||
_session_label_c = "#DAA520"
|
||
_session_border_c = "#8B8682"
|
||
_assistant_label_c = "#8FBC8F"
|
||
|
||
lines = Text()
|
||
if skipped:
|
||
lines.append(
|
||
f" ... {skipped} earlier messages ...\n\n",
|
||
style="dim italic",
|
||
)
|
||
|
||
for i, (role, text) in enumerate(entries):
|
||
if role == "user":
|
||
lines.append(" ● You: ", style=f"dim bold {_session_label_c}")
|
||
# Show first line inline, indent rest
|
||
msg_lines = text.splitlines()
|
||
lines.append(msg_lines[0] + "\n", style="dim")
|
||
for ml in msg_lines[1:]:
|
||
lines.append(f" {ml}\n", style="dim")
|
||
else:
|
||
lines.append(" ◆ Hermes: ", style=f"dim bold {_assistant_label_c}")
|
||
msg_lines = text.splitlines()
|
||
lines.append(msg_lines[0] + "\n", style="dim")
|
||
for ml in msg_lines[1:]:
|
||
lines.append(f" {ml}\n", style="dim")
|
||
if i < len(entries) - 1:
|
||
lines.append("") # small gap
|
||
|
||
panel = Panel(
|
||
lines,
|
||
title=f"[dim {_session_label_c}]Previous Conversation[/]",
|
||
border_style=f"dim {_session_border_c}",
|
||
padding=(0, 1),
|
||
style=_history_text_c,
|
||
)
|
||
self.console.print(panel)
|
||
|
||
def _try_attach_clipboard_image(self) -> bool:
|
||
"""Check clipboard for an image and attach it if found.
|
||
|
||
Saves the image to ~/.hermes/images/ and appends the path to
|
||
``_attached_images``. Returns True if an image was attached.
|
||
"""
|
||
from hermes_cli.clipboard import save_clipboard_image
|
||
|
||
img_dir = Path(os.getenv("HERMES_HOME", Path.home() / ".hermes")) / "images"
|
||
self._image_counter += 1
|
||
ts = datetime.now().strftime("%Y%m%d_%H%M%S")
|
||
img_path = img_dir / f"clip_{ts}_{self._image_counter}.png"
|
||
|
||
if save_clipboard_image(img_path):
|
||
self._attached_images.append(img_path)
|
||
return True
|
||
self._image_counter -= 1
|
||
return False
|
||
|
||
def _handle_rollback_command(self, command: str):
|
||
"""Handle /rollback — list, diff, or restore filesystem checkpoints.
|
||
|
||
Syntax:
|
||
/rollback — list checkpoints
|
||
/rollback <N> — restore checkpoint N (also undoes last chat turn)
|
||
/rollback diff <N> — preview changes since checkpoint N
|
||
/rollback <N> <file> — restore a single file from checkpoint N
|
||
"""
|
||
from tools.checkpoint_manager import CheckpointManager, format_checkpoint_list
|
||
|
||
if not hasattr(self, 'agent') or not self.agent:
|
||
print(" No active agent session.")
|
||
return
|
||
|
||
mgr = self.agent._checkpoint_mgr
|
||
if not mgr.enabled:
|
||
print(" Checkpoints are not enabled.")
|
||
print(" Enable with: hermes --checkpoints")
|
||
print(" Or in config.yaml: checkpoints: { enabled: true }")
|
||
return
|
||
|
||
cwd = os.getenv("TERMINAL_CWD", os.getcwd())
|
||
parts = command.split()
|
||
args = parts[1:] if len(parts) > 1 else []
|
||
|
||
if not args:
|
||
# List checkpoints
|
||
checkpoints = mgr.list_checkpoints(cwd)
|
||
print(format_checkpoint_list(checkpoints, cwd))
|
||
return
|
||
|
||
# Handle /rollback diff <N>
|
||
if args[0].lower() == "diff":
|
||
if len(args) < 2:
|
||
print(" Usage: /rollback diff <N>")
|
||
return
|
||
checkpoints = mgr.list_checkpoints(cwd)
|
||
if not checkpoints:
|
||
print(f" No checkpoints found for {cwd}")
|
||
return
|
||
target_hash = self._resolve_checkpoint_ref(args[1], checkpoints)
|
||
if not target_hash:
|
||
return
|
||
result = mgr.diff(cwd, target_hash)
|
||
if result["success"]:
|
||
stat = result.get("stat", "")
|
||
diff = result.get("diff", "")
|
||
if not stat and not diff:
|
||
print(" No changes since this checkpoint.")
|
||
else:
|
||
if stat:
|
||
print(f"\n{stat}")
|
||
if diff:
|
||
# Limit diff output to avoid terminal flood
|
||
diff_lines = diff.splitlines()
|
||
if len(diff_lines) > 80:
|
||
print("\n".join(diff_lines[:80]))
|
||
print(f"\n ... ({len(diff_lines) - 80} more lines, showing first 80)")
|
||
else:
|
||
print(f"\n{diff}")
|
||
else:
|
||
print(f" ❌ {result['error']}")
|
||
return
|
||
|
||
# Resolve checkpoint reference (number or hash)
|
||
checkpoints = mgr.list_checkpoints(cwd)
|
||
if not checkpoints:
|
||
print(f" No checkpoints found for {cwd}")
|
||
return
|
||
|
||
target_hash = self._resolve_checkpoint_ref(args[0], checkpoints)
|
||
if not target_hash:
|
||
return
|
||
|
||
# Check for file-level restore: /rollback <N> <file>
|
||
file_path = args[1] if len(args) > 1 else None
|
||
|
||
result = mgr.restore(cwd, target_hash, file_path=file_path)
|
||
if result["success"]:
|
||
if file_path:
|
||
print(f" ✅ Restored {file_path} from checkpoint {result['restored_to']}: {result['reason']}")
|
||
else:
|
||
print(f" ✅ Restored to checkpoint {result['restored_to']}: {result['reason']}")
|
||
print(f" A pre-rollback snapshot was saved automatically.")
|
||
|
||
# Also undo the last conversation turn so the agent's context
|
||
# matches the restored filesystem state
|
||
if self.conversation_history:
|
||
self.undo_last()
|
||
print(f" Chat turn undone to match restored file state.")
|
||
else:
|
||
print(f" ❌ {result['error']}")
|
||
|
||
def _resolve_checkpoint_ref(self, ref: str, checkpoints: list) -> str | None:
|
||
"""Resolve a checkpoint number or hash to a full commit hash."""
|
||
try:
|
||
idx = int(ref) - 1 # 1-indexed for user
|
||
if 0 <= idx < len(checkpoints):
|
||
return checkpoints[idx]["hash"]
|
||
else:
|
||
print(f" Invalid checkpoint number. Use 1-{len(checkpoints)}.")
|
||
return None
|
||
except ValueError:
|
||
# Treat as a git hash
|
||
return ref
|
||
|
||
def _handle_stop_command(self):
|
||
"""Handle /stop — kill all running background processes.
|
||
|
||
Inspired by OpenAI Codex's separation of interrupt (stop current turn)
|
||
from /stop (clean up background processes). See openai/codex#14602.
|
||
"""
|
||
from tools.process_registry import get_registry
|
||
|
||
registry = get_registry()
|
||
processes = registry.list_processes()
|
||
running = [p for p in processes if p.get("status") == "running"]
|
||
|
||
if not running:
|
||
print(" No running background processes.")
|
||
return
|
||
|
||
print(f" Stopping {len(running)} background process(es)...")
|
||
killed = registry.kill_all()
|
||
print(f" ✅ Stopped {killed} process(es).")
|
||
|
||
def _handle_paste_command(self):
|
||
"""Handle /paste — explicitly check clipboard for an image.
|
||
|
||
This is the reliable fallback for terminals where BracketedPaste
|
||
doesn't fire for image-only clipboard content (e.g., VSCode terminal,
|
||
Windows Terminal with WSL2).
|
||
"""
|
||
from hermes_cli.clipboard import has_clipboard_image
|
||
if has_clipboard_image():
|
||
if self._try_attach_clipboard_image():
|
||
n = len(self._attached_images)
|
||
_cprint(f" 📎 Image #{n} attached from clipboard")
|
||
else:
|
||
_cprint(f" {_DIM}(>_<) Clipboard has an image but extraction failed{_RST}")
|
||
else:
|
||
_cprint(f" {_DIM}(._.) No image found in clipboard{_RST}")
|
||
|
||
def _preprocess_images_with_vision(self, text: str, images: list) -> str:
|
||
"""Analyze attached images via the vision tool and return enriched text.
|
||
|
||
Instead of embedding raw base64 ``image_url`` content parts in the
|
||
conversation (which only works with vision-capable models), this
|
||
pre-processes each image through the auxiliary vision model (Gemini
|
||
Flash) and prepends the descriptions to the user's message — the
|
||
same approach the messaging gateway uses.
|
||
|
||
The local file path is included so the agent can re-examine the
|
||
image later with ``vision_analyze`` if needed.
|
||
"""
|
||
import asyncio as _asyncio
|
||
import json as _json
|
||
from tools.vision_tools import vision_analyze_tool
|
||
|
||
analysis_prompt = (
|
||
"Describe everything visible in this image in thorough detail. "
|
||
"Include any text, code, data, objects, people, layout, colors, "
|
||
"and any other notable visual information."
|
||
)
|
||
|
||
enriched_parts = []
|
||
for img_path in images:
|
||
if not img_path.exists():
|
||
continue
|
||
size_kb = img_path.stat().st_size // 1024
|
||
_cprint(f" {_DIM}👁️ analyzing {img_path.name} ({size_kb}KB)...{_RST}")
|
||
try:
|
||
result_json = _asyncio.run(
|
||
vision_analyze_tool(image_url=str(img_path), user_prompt=analysis_prompt)
|
||
)
|
||
result = _json.loads(result_json)
|
||
if result.get("success"):
|
||
description = result.get("analysis", "")
|
||
enriched_parts.append(
|
||
f"[The user attached an image. Here's what it contains:\n{description}]\n"
|
||
f"[If you need a closer look, use vision_analyze with "
|
||
f"image_url: {img_path}]"
|
||
)
|
||
_cprint(f" {_DIM}✓ image analyzed{_RST}")
|
||
else:
|
||
enriched_parts.append(
|
||
f"[The user attached an image but it couldn't be analyzed. "
|
||
f"You can try examining it with vision_analyze using "
|
||
f"image_url: {img_path}]"
|
||
)
|
||
_cprint(f" {_DIM}⚠ vision analysis failed — path included for retry{_RST}")
|
||
except Exception as e:
|
||
enriched_parts.append(
|
||
f"[The user attached an image but analysis failed ({e}). "
|
||
f"You can try examining it with vision_analyze using "
|
||
f"image_url: {img_path}]"
|
||
)
|
||
_cprint(f" {_DIM}⚠ vision analysis error — path included for retry{_RST}")
|
||
|
||
# Combine: vision descriptions first, then the user's original text
|
||
user_text = text if isinstance(text, str) and text else ""
|
||
if enriched_parts:
|
||
prefix = "\n\n".join(enriched_parts)
|
||
return f"{prefix}\n\n{user_text}" if user_text else prefix
|
||
return user_text or "What do you see in this image?"
|
||
|
||
def _show_tool_availability_warnings(self):
|
||
"""Show warnings about disabled tools due to missing API keys."""
|
||
try:
|
||
from model_tools import check_tool_availability, TOOLSET_REQUIREMENTS
|
||
|
||
available, unavailable = check_tool_availability()
|
||
|
||
# Filter to only those missing API keys (not system deps)
|
||
api_key_missing = [u for u in unavailable if u["missing_vars"]]
|
||
|
||
if api_key_missing:
|
||
self.console.print()
|
||
self.console.print("[yellow]⚠️ Some tools disabled (missing API keys):[/]")
|
||
for item in api_key_missing:
|
||
tools_str = ", ".join(item["tools"][:2]) # Show first 2 tools
|
||
if len(item["tools"]) > 2:
|
||
tools_str += f", +{len(item['tools'])-2} more"
|
||
self.console.print(f" [dim]• {item['name']}[/] [dim italic]({', '.join(item['missing_vars'])})[/]")
|
||
self.console.print("[dim] Run 'hermes setup' to configure[/]")
|
||
except Exception:
|
||
pass # Don't crash on import errors
|
||
|
||
def _show_status(self):
|
||
"""Show current status bar."""
|
||
# Get tool count
|
||
tools = get_tool_definitions(enabled_toolsets=self.enabled_toolsets, quiet_mode=True)
|
||
tool_count = len(tools) if tools else 0
|
||
|
||
# Format model name (shorten if needed)
|
||
model_short = self.model.split("/")[-1] if "/" in self.model else self.model
|
||
if len(model_short) > 30:
|
||
model_short = model_short[:27] + "..."
|
||
|
||
# Get API status indicator
|
||
if self.api_key:
|
||
api_indicator = "[green bold]●[/]"
|
||
else:
|
||
api_indicator = "[red bold]●[/]"
|
||
|
||
# Build status line with proper markup
|
||
toolsets_info = ""
|
||
if self.enabled_toolsets and "all" not in self.enabled_toolsets:
|
||
toolsets_info = f" [dim #B8860B]·[/] [#CD7F32]toolsets: {', '.join(self.enabled_toolsets)}[/]"
|
||
|
||
provider_info = f" [dim #B8860B]·[/] [dim]provider: {self.provider}[/]"
|
||
if self._provider_source:
|
||
provider_info += f" [dim #B8860B]·[/] [dim]auth: {self._provider_source}[/]"
|
||
|
||
self.console.print(
|
||
f" {api_indicator} [#FFBF00]{model_short}[/] "
|
||
f"[dim #B8860B]·[/] [bold cyan]{tool_count} tools[/]"
|
||
f"{toolsets_info}{provider_info}"
|
||
)
|
||
|
||
def show_help(self):
|
||
"""Display help information with categorized commands."""
|
||
from hermes_cli.commands import COMMANDS_BY_CATEGORY
|
||
|
||
try:
|
||
from hermes_cli.skin_engine import get_active_help_header
|
||
header = get_active_help_header("(^_^)? Available Commands")
|
||
except Exception:
|
||
header = "(^_^)? Available Commands"
|
||
header = (header or "").strip() or "(^_^)? Available Commands"
|
||
inner_width = 55
|
||
if len(header) > inner_width:
|
||
header = header[:inner_width]
|
||
_cprint(f"\n{_BOLD}+{'-' * inner_width}+{_RST}")
|
||
_cprint(f"{_BOLD}|{header:^{inner_width}}|{_RST}")
|
||
_cprint(f"{_BOLD}+{'-' * inner_width}+{_RST}")
|
||
|
||
for category, commands in COMMANDS_BY_CATEGORY.items():
|
||
_cprint(f"\n {_BOLD}── {category} ──{_RST}")
|
||
for cmd, desc in commands.items():
|
||
ChatConsole().print(f" [bold {_accent_hex()}]{cmd:<15}[/] [dim]-[/] {_escape(desc)}")
|
||
|
||
if _skill_commands:
|
||
_cprint(f"\n ⚡ {_BOLD}Skill Commands{_RST} ({len(_skill_commands)} installed):")
|
||
for cmd, info in sorted(_skill_commands.items()):
|
||
ChatConsole().print(
|
||
f" [bold {_accent_hex()}]{cmd:<22}[/] [dim]-[/] {_escape(info['description'])}"
|
||
)
|
||
|
||
_cprint(f"\n {_DIM}Tip: Just type your message to chat with Hermes!{_RST}")
|
||
_cprint(f" {_DIM}Multi-line: Alt+Enter for a new line{_RST}")
|
||
_cprint(f" {_DIM}Paste image: Alt+V (or /paste){_RST}\n")
|
||
|
||
def show_tools(self):
|
||
"""Display available tools with kawaii ASCII art."""
|
||
tools = get_tool_definitions(enabled_toolsets=self.enabled_toolsets, quiet_mode=True)
|
||
|
||
if not tools:
|
||
print("(;_;) No tools available")
|
||
return
|
||
|
||
# Header
|
||
print()
|
||
title = "(^_^)/ Available Tools"
|
||
width = 78
|
||
pad = width - len(title)
|
||
print("+" + "-" * width + "+")
|
||
print("|" + " " * (pad // 2) + title + " " * (pad - pad // 2) + "|")
|
||
print("+" + "-" * width + "+")
|
||
print()
|
||
|
||
# Group tools by toolset
|
||
toolsets = {}
|
||
for tool in sorted(tools, key=lambda t: t["function"]["name"]):
|
||
name = tool["function"]["name"]
|
||
toolset = get_toolset_for_tool(name) or "unknown"
|
||
if toolset not in toolsets:
|
||
toolsets[toolset] = []
|
||
desc = tool["function"].get("description", "")
|
||
# First sentence: split on ". " (period+space) to avoid breaking on "e.g." or "v2.0"
|
||
desc = desc.split("\n")[0]
|
||
if ". " in desc:
|
||
desc = desc[:desc.index(". ") + 1]
|
||
toolsets[toolset].append((name, desc))
|
||
|
||
# Display by toolset
|
||
for toolset in sorted(toolsets.keys()):
|
||
print(f" [{toolset}]")
|
||
for name, desc in toolsets[toolset]:
|
||
print(f" * {name:<20} - {desc}")
|
||
print()
|
||
|
||
print(f" Total: {len(tools)} tools ヽ(^o^)ノ")
|
||
print()
|
||
|
||
def _handle_tools_command(self, cmd: str):
|
||
"""Handle /tools [list|disable|enable] slash commands.
|
||
|
||
/tools (no args) shows the tool list.
|
||
/tools list shows enabled/disabled status per toolset.
|
||
/tools disable/enable saves the change to config and resets
|
||
the session so the new tool set takes effect cleanly (no
|
||
prompt-cache breakage mid-conversation).
|
||
"""
|
||
import shlex
|
||
from argparse import Namespace
|
||
from hermes_cli.tools_config import tools_disable_enable_command
|
||
|
||
try:
|
||
parts = shlex.split(cmd)
|
||
except ValueError:
|
||
parts = cmd.split()
|
||
|
||
subcommand = parts[1] if len(parts) > 1 else ""
|
||
if subcommand not in ("list", "disable", "enable"):
|
||
self.show_tools()
|
||
return
|
||
|
||
if subcommand == "list":
|
||
tools_disable_enable_command(
|
||
Namespace(tools_action="list", platform="cli"))
|
||
return
|
||
|
||
names = parts[2:]
|
||
if not names:
|
||
print(f"(._.) Usage: /tools {subcommand} <name> [name ...]")
|
||
print(f" Built-in toolset: /tools {subcommand} web")
|
||
print(f" MCP tool: /tools {subcommand} github:create_issue")
|
||
return
|
||
|
||
# Confirm session reset before applying
|
||
verb = "Disable" if subcommand == "disable" else "Enable"
|
||
label = ", ".join(names)
|
||
_cprint(f"{_GOLD}{verb} {label}?{_RST}")
|
||
_cprint(f"{_DIM}This will save to config and reset your session so the "
|
||
f"change takes effect cleanly.{_RST}")
|
||
try:
|
||
answer = input(" Continue? [y/N] ").strip().lower()
|
||
except (EOFError, KeyboardInterrupt):
|
||
print()
|
||
_cprint(f"{_DIM}Cancelled.{_RST}")
|
||
return
|
||
|
||
if answer not in ("y", "yes"):
|
||
_cprint(f"{_DIM}Cancelled.{_RST}")
|
||
return
|
||
|
||
tools_disable_enable_command(
|
||
Namespace(tools_action=subcommand, names=names, platform="cli"))
|
||
|
||
# Reset session so the new tool config is picked up from a clean state
|
||
from hermes_cli.tools_config import _get_platform_tools
|
||
from hermes_cli.config import load_config
|
||
self.enabled_toolsets = _get_platform_tools(load_config(), "cli")
|
||
self.new_session()
|
||
_cprint(f"{_DIM}Session reset. New tool configuration is active.{_RST}")
|
||
|
||
def show_toolsets(self):
|
||
"""Display available toolsets with kawaii ASCII art."""
|
||
all_toolsets = get_all_toolsets()
|
||
|
||
# Header
|
||
print()
|
||
title = "(^_^)b Available Toolsets"
|
||
width = 58
|
||
pad = width - len(title)
|
||
print("+" + "-" * width + "+")
|
||
print("|" + " " * (pad // 2) + title + " " * (pad - pad // 2) + "|")
|
||
print("+" + "-" * width + "+")
|
||
print()
|
||
|
||
for name in sorted(all_toolsets.keys()):
|
||
info = get_toolset_info(name)
|
||
if info:
|
||
tool_count = info["tool_count"]
|
||
desc = info["description"]
|
||
|
||
# Mark if currently enabled
|
||
marker = "(*)" if self.enabled_toolsets and name in self.enabled_toolsets else " "
|
||
print(f" {marker} {name:<18} [{tool_count:>2} tools] - {desc}")
|
||
|
||
print()
|
||
print(" (*) = currently enabled")
|
||
print()
|
||
print(" Tip: Use 'all' or '*' to enable all toolsets")
|
||
print(" Example: python cli.py --toolsets web,terminal")
|
||
print()
|
||
|
||
def show_config(self):
|
||
"""Display current configuration with kawaii ASCII art."""
|
||
# Get terminal config from environment (which was set from cli-config.yaml)
|
||
terminal_env = os.getenv("TERMINAL_ENV", "local")
|
||
terminal_cwd = os.getenv("TERMINAL_CWD", os.getcwd())
|
||
terminal_timeout = os.getenv("TERMINAL_TIMEOUT", "60")
|
||
|
||
user_config_path = _hermes_home / 'config.yaml'
|
||
project_config_path = Path(__file__).parent / 'cli-config.yaml'
|
||
if user_config_path.exists():
|
||
config_path = user_config_path
|
||
else:
|
||
config_path = project_config_path
|
||
config_status = "(loaded)" if config_path.exists() else "(not found)"
|
||
|
||
api_key_display = '********' + self.api_key[-4:] if self.api_key and len(self.api_key) > 4 else 'Not set!'
|
||
|
||
print()
|
||
title = "(^_^) Configuration"
|
||
width = 50
|
||
pad = width - len(title)
|
||
print("+" + "-" * width + "+")
|
||
print("|" + " " * (pad // 2) + title + " " * (pad - pad // 2) + "|")
|
||
print("+" + "-" * width + "+")
|
||
print()
|
||
print(" -- Model --")
|
||
print(f" Model: {self.model}")
|
||
print(f" Base URL: {self.base_url}")
|
||
print(f" API Key: {api_key_display}")
|
||
print()
|
||
print(" -- Terminal --")
|
||
print(f" Environment: {terminal_env}")
|
||
if terminal_env == "ssh":
|
||
ssh_host = os.getenv("TERMINAL_SSH_HOST", "not set")
|
||
ssh_user = os.getenv("TERMINAL_SSH_USER", "not set")
|
||
ssh_port = os.getenv("TERMINAL_SSH_PORT", "22")
|
||
print(f" SSH Target: {ssh_user}@{ssh_host}:{ssh_port}")
|
||
print(f" Working Dir: {terminal_cwd}")
|
||
print(f" Timeout: {terminal_timeout}s")
|
||
print()
|
||
print(" -- Agent --")
|
||
print(f" Max Turns: {self.max_turns}")
|
||
print(f" Toolsets: {', '.join(self.enabled_toolsets) if self.enabled_toolsets else 'all'}")
|
||
print(f" Verbose: {self.verbose}")
|
||
print()
|
||
print(" -- Session --")
|
||
print(f" Started: {self.session_start.strftime('%Y-%m-%d %H:%M:%S')}")
|
||
print(f" Config File: {config_path} {config_status}")
|
||
print()
|
||
|
||
def show_history(self):
|
||
"""Display conversation history."""
|
||
if not self.conversation_history:
|
||
print("(._.) No conversation history yet.")
|
||
return
|
||
|
||
preview_limit = 400
|
||
visible_index = 0
|
||
hidden_tool_messages = 0
|
||
|
||
def flush_tool_summary():
|
||
nonlocal hidden_tool_messages
|
||
if not hidden_tool_messages:
|
||
return
|
||
|
||
noun = "message" if hidden_tool_messages == 1 else "messages"
|
||
print("\n [Tools]")
|
||
print(f" ({hidden_tool_messages} tool {noun} hidden)")
|
||
hidden_tool_messages = 0
|
||
|
||
print()
|
||
print("+" + "-" * 50 + "+")
|
||
print("|" + " " * 12 + "(^_^) Conversation History" + " " * 11 + "|")
|
||
print("+" + "-" * 50 + "+")
|
||
|
||
for msg in self.conversation_history:
|
||
role = msg.get("role", "unknown")
|
||
|
||
if role == "tool":
|
||
hidden_tool_messages += 1
|
||
continue
|
||
|
||
if role not in {"user", "assistant"}:
|
||
continue
|
||
|
||
flush_tool_summary()
|
||
visible_index += 1
|
||
|
||
content = msg.get("content")
|
||
content_text = "" if content is None else str(content)
|
||
|
||
if role == "user":
|
||
print(f"\n [You #{visible_index}]")
|
||
print(
|
||
f" {content_text[:preview_limit]}{'...' if len(content_text) > preview_limit else ''}"
|
||
)
|
||
continue
|
||
|
||
print(f"\n [Hermes #{visible_index}]")
|
||
tool_calls = msg.get("tool_calls") or []
|
||
if content_text:
|
||
preview = content_text[:preview_limit]
|
||
suffix = "..." if len(content_text) > preview_limit else ""
|
||
elif tool_calls:
|
||
tool_count = len(tool_calls)
|
||
noun = "call" if tool_count == 1 else "calls"
|
||
preview = f"(requested {tool_count} tool {noun})"
|
||
suffix = ""
|
||
else:
|
||
preview = "(no text response)"
|
||
suffix = ""
|
||
print(f" {preview}{suffix}")
|
||
|
||
flush_tool_summary()
|
||
print()
|
||
|
||
def new_session(self, silent=False):
|
||
"""Start a fresh session with a new session ID and cleared agent state."""
|
||
if self.agent and self.conversation_history:
|
||
try:
|
||
self.agent.flush_memories(self.conversation_history)
|
||
except Exception:
|
||
pass
|
||
|
||
old_session_id = self.session_id
|
||
if self._session_db and old_session_id:
|
||
try:
|
||
self._session_db.end_session(old_session_id, "new_session")
|
||
except Exception:
|
||
pass
|
||
|
||
self.session_start = datetime.now()
|
||
timestamp_str = self.session_start.strftime("%Y%m%d_%H%M%S")
|
||
short_uuid = uuid.uuid4().hex[:6]
|
||
self.session_id = f"{timestamp_str}_{short_uuid}"
|
||
self.conversation_history = []
|
||
self._pending_title = None
|
||
self._resumed = False
|
||
|
||
if self.agent:
|
||
self.agent.session_id = self.session_id
|
||
self.agent.session_start = self.session_start
|
||
self.agent.reset_session_state()
|
||
if hasattr(self.agent, "_last_flushed_db_idx"):
|
||
self.agent._last_flushed_db_idx = 0
|
||
if hasattr(self.agent, "_todo_store"):
|
||
try:
|
||
from tools.todo_tool import TodoStore
|
||
self.agent._todo_store = TodoStore()
|
||
except Exception:
|
||
pass
|
||
if hasattr(self.agent, "_invalidate_system_prompt"):
|
||
self.agent._invalidate_system_prompt()
|
||
|
||
if self._session_db:
|
||
try:
|
||
self._session_db.create_session(
|
||
session_id=self.session_id,
|
||
source="cli",
|
||
model=self.model,
|
||
model_config={
|
||
"max_iterations": self.max_turns,
|
||
"reasoning_config": self.reasoning_config,
|
||
},
|
||
)
|
||
except Exception:
|
||
pass
|
||
|
||
if not silent:
|
||
print("(^_^)v New session started!")
|
||
|
||
def reset_conversation(self):
|
||
"""Reset the conversation by starting a new session."""
|
||
self.new_session()
|
||
|
||
def save_conversation(self):
|
||
"""Save the current conversation to a file."""
|
||
if not self.conversation_history:
|
||
print("(;_;) No conversation to save.")
|
||
return
|
||
|
||
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
|
||
filename = f"hermes_conversation_{timestamp}.json"
|
||
|
||
try:
|
||
with open(filename, "w", encoding="utf-8") as f:
|
||
json.dump({
|
||
"model": self.model,
|
||
"session_start": self.session_start.isoformat(),
|
||
"messages": self.conversation_history,
|
||
}, f, indent=2, ensure_ascii=False)
|
||
print(f"(^_^)v Conversation saved to: {filename}")
|
||
except Exception as e:
|
||
print(f"(x_x) Failed to save: {e}")
|
||
|
||
def retry_last(self):
|
||
"""Retry the last user message by removing the last exchange and re-sending.
|
||
|
||
Removes the last assistant response (and any tool-call messages) and
|
||
the last user message, then re-sends that user message to the agent.
|
||
Returns the message to re-send, or None if there's nothing to retry.
|
||
"""
|
||
if not self.conversation_history:
|
||
print("(._.) No messages to retry.")
|
||
return None
|
||
|
||
# Walk backwards to find the last user message
|
||
last_user_idx = None
|
||
for i in range(len(self.conversation_history) - 1, -1, -1):
|
||
if self.conversation_history[i].get("role") == "user":
|
||
last_user_idx = i
|
||
break
|
||
|
||
if last_user_idx is None:
|
||
print("(._.) No user message found to retry.")
|
||
return None
|
||
|
||
# Extract the message text and remove everything from that point forward
|
||
last_message = self.conversation_history[last_user_idx].get("content", "")
|
||
self.conversation_history = self.conversation_history[:last_user_idx]
|
||
|
||
print(f"(^_^)b Retrying: \"{last_message[:60]}{'...' if len(last_message) > 60 else ''}\"")
|
||
return last_message
|
||
|
||
def undo_last(self):
|
||
"""Remove the last user/assistant exchange from conversation history.
|
||
|
||
Walks backwards and removes all messages from the last user message
|
||
onward (including assistant responses, tool calls, etc.).
|
||
"""
|
||
if not self.conversation_history:
|
||
print("(._.) No messages to undo.")
|
||
return
|
||
|
||
# Walk backwards to find the last user message
|
||
last_user_idx = None
|
||
for i in range(len(self.conversation_history) - 1, -1, -1):
|
||
if self.conversation_history[i].get("role") == "user":
|
||
last_user_idx = i
|
||
break
|
||
|
||
if last_user_idx is None:
|
||
print("(._.) No user message found to undo.")
|
||
return
|
||
|
||
# Count how many messages we're removing
|
||
removed_count = len(self.conversation_history) - last_user_idx
|
||
removed_msg = self.conversation_history[last_user_idx].get("content", "")
|
||
|
||
# Truncate history to before the last user message
|
||
self.conversation_history = self.conversation_history[:last_user_idx]
|
||
|
||
print(f"(^_^)b Undid {removed_count} message(s). Removed: \"{removed_msg[:60]}{'...' if len(removed_msg) > 60 else ''}\"")
|
||
remaining = len(self.conversation_history)
|
||
print(f" {remaining} message(s) remaining in history.")
|
||
|
||
def _show_model_and_providers(self):
|
||
"""Unified /model and /provider display.
|
||
|
||
Shows current model + provider, then lists all authenticated
|
||
providers with their available models so users can switch easily.
|
||
"""
|
||
from hermes_cli.models import (
|
||
curated_models_for_provider, list_available_providers,
|
||
normalize_provider, _PROVIDER_LABELS,
|
||
)
|
||
from hermes_cli.auth import resolve_provider as _resolve_provider
|
||
|
||
# Resolve current provider
|
||
raw_provider = normalize_provider(self.provider)
|
||
if raw_provider == "auto":
|
||
try:
|
||
current = _resolve_provider(
|
||
self.requested_provider,
|
||
explicit_api_key=self._explicit_api_key,
|
||
explicit_base_url=self._explicit_base_url,
|
||
)
|
||
except Exception:
|
||
current = "openrouter"
|
||
else:
|
||
current = raw_provider
|
||
current_label = _PROVIDER_LABELS.get(current, current)
|
||
|
||
print(f"\n Current: {self.model} via {current_label}")
|
||
print()
|
||
|
||
# Show all authenticated providers with their models
|
||
providers = list_available_providers()
|
||
authed = [p for p in providers if p["authenticated"]]
|
||
unauthed = [p for p in providers if not p["authenticated"]]
|
||
|
||
if authed:
|
||
print(" Authenticated providers & models:")
|
||
for p in authed:
|
||
is_active = p["id"] == current
|
||
marker = " ← active" if is_active else ""
|
||
print(f" [{p['id']}]{marker}")
|
||
curated = curated_models_for_provider(p["id"])
|
||
if curated:
|
||
for mid, desc in curated:
|
||
current_marker = " ← current" if (is_active and mid == self.model) else ""
|
||
print(f" {mid}{current_marker}")
|
||
elif p["id"] == "custom":
|
||
from hermes_cli.models import _get_custom_base_url
|
||
custom_url = _get_custom_base_url() or os.getenv("OPENAI_BASE_URL", "")
|
||
if custom_url:
|
||
print(f" endpoint: {custom_url}")
|
||
if is_active:
|
||
print(f" model: {self.model} ← current")
|
||
print(f" (use /model custom:<model-name>)")
|
||
else:
|
||
print(f" (use /model {p['id']}:<model-name>)")
|
||
print()
|
||
|
||
if unauthed:
|
||
names = ", ".join(p["label"] for p in unauthed)
|
||
print(f" Not configured: {names}")
|
||
print(f" Run: hermes setup")
|
||
print()
|
||
|
||
print(" Switch model: /model <model-name>")
|
||
print(" Switch provider: /model <provider>:<model-name>")
|
||
if authed and len(authed) > 1:
|
||
# Show a concrete example with a non-active provider
|
||
other = next((p for p in authed if p["id"] != current), authed[0])
|
||
other_models = curated_models_for_provider(other["id"])
|
||
if other_models:
|
||
example_model = other_models[0][0]
|
||
print(f" Example: /model {other['id']}:{example_model}")
|
||
|
||
def _handle_prompt_command(self, cmd: str):
|
||
"""Handle the /prompt command to view or set system prompt."""
|
||
parts = cmd.split(maxsplit=1)
|
||
|
||
if len(parts) > 1:
|
||
# Set new prompt
|
||
new_prompt = parts[1].strip()
|
||
|
||
if new_prompt.lower() == "clear":
|
||
self.system_prompt = ""
|
||
self.agent = None # Force re-init
|
||
if save_config_value("agent.system_prompt", ""):
|
||
print("(^_^)b System prompt cleared (saved to config)")
|
||
else:
|
||
print("(^_^) System prompt cleared (session only)")
|
||
else:
|
||
self.system_prompt = new_prompt
|
||
self.agent = None # Force re-init
|
||
if save_config_value("agent.system_prompt", new_prompt):
|
||
print(f"(^_^)b System prompt set (saved to config)")
|
||
else:
|
||
print(f"(^_^) System prompt set (session only)")
|
||
print(f" \"{new_prompt[:60]}{'...' if len(new_prompt) > 60 else ''}\"")
|
||
else:
|
||
# Show current prompt
|
||
print()
|
||
print("+" + "-" * 50 + "+")
|
||
print("|" + " " * 15 + "(^_^) System Prompt" + " " * 15 + "|")
|
||
print("+" + "-" * 50 + "+")
|
||
print()
|
||
if self.system_prompt:
|
||
# Word wrap the prompt for display
|
||
words = self.system_prompt.split()
|
||
lines = []
|
||
current_line = ""
|
||
for word in words:
|
||
if len(current_line) + len(word) + 1 <= 50:
|
||
current_line += (" " if current_line else "") + word
|
||
else:
|
||
lines.append(current_line)
|
||
current_line = word
|
||
if current_line:
|
||
lines.append(current_line)
|
||
for line in lines:
|
||
print(f" {line}")
|
||
else:
|
||
print(" (no custom prompt set - using default)")
|
||
print()
|
||
print(" Usage:")
|
||
print(" /prompt <text> - Set a custom system prompt")
|
||
print(" /prompt clear - Remove custom prompt")
|
||
print(" /personality - Use a predefined personality")
|
||
print()
|
||
|
||
|
||
@staticmethod
|
||
def _resolve_personality_prompt(value) -> str:
|
||
"""Accept string or dict personality value; return system prompt string."""
|
||
if isinstance(value, dict):
|
||
parts = [value.get("system_prompt", "")]
|
||
if value.get("tone"):
|
||
parts.append(f'Tone: {value["tone"]}' )
|
||
if value.get("style"):
|
||
parts.append(f'Style: {value["style"]}' )
|
||
return "\n".join(p for p in parts if p)
|
||
return str(value)
|
||
|
||
def _handle_personality_command(self, cmd: str):
|
||
"""Handle the /personality command to set predefined personalities."""
|
||
parts = cmd.split(maxsplit=1)
|
||
|
||
if len(parts) > 1:
|
||
# Set personality
|
||
personality_name = parts[1].strip().lower()
|
||
|
||
if personality_name in ("none", "default", "neutral"):
|
||
self.system_prompt = ""
|
||
self.agent = None # Force re-init
|
||
if save_config_value("agent.system_prompt", ""):
|
||
print("(^_^)b Personality cleared (saved to config)")
|
||
else:
|
||
print("(^_^) Personality cleared (session only)")
|
||
print(" No personality overlay — using base agent behavior.")
|
||
elif personality_name in self.personalities:
|
||
self.system_prompt = self._resolve_personality_prompt(self.personalities[personality_name])
|
||
self.agent = None # Force re-init
|
||
if save_config_value("agent.system_prompt", self.system_prompt):
|
||
print(f"(^_^)b Personality set to '{personality_name}' (saved to config)")
|
||
else:
|
||
print(f"(^_^) Personality set to '{personality_name}' (session only)")
|
||
print(f" \"{self.system_prompt[:60]}{'...' if len(self.system_prompt) > 60 else ''}\"")
|
||
else:
|
||
print(f"(._.) Unknown personality: {personality_name}")
|
||
print(f" Available: none, {', '.join(self.personalities.keys())}")
|
||
else:
|
||
# Show available personalities
|
||
print()
|
||
print("+" + "-" * 50 + "+")
|
||
print("|" + " " * 12 + "(^o^)/ Personalities" + " " * 15 + "|")
|
||
print("+" + "-" * 50 + "+")
|
||
print()
|
||
print(f" {'none':<12} - (no personality overlay)")
|
||
for name, prompt in self.personalities.items():
|
||
if isinstance(prompt, dict):
|
||
preview = prompt.get("description") or prompt.get("system_prompt", "")[:50]
|
||
else:
|
||
preview = str(prompt)[:50]
|
||
print(f" {name:<12} - {preview}")
|
||
print()
|
||
print(" Usage: /personality <name>")
|
||
print()
|
||
|
||
def _handle_cron_command(self, cmd: str):
|
||
"""Handle the /cron command to manage scheduled tasks."""
|
||
import shlex
|
||
from tools.cronjob_tools import cronjob as cronjob_tool
|
||
|
||
def _cron_api(**kwargs):
|
||
return json.loads(cronjob_tool(**kwargs))
|
||
|
||
def _normalize_skills(values):
|
||
normalized = []
|
||
for value in values:
|
||
text = str(value or "").strip()
|
||
if text and text not in normalized:
|
||
normalized.append(text)
|
||
return normalized
|
||
|
||
def _parse_flags(tokens):
|
||
opts = {
|
||
"name": None,
|
||
"deliver": None,
|
||
"repeat": None,
|
||
"skills": [],
|
||
"add_skills": [],
|
||
"remove_skills": [],
|
||
"clear_skills": False,
|
||
"all": False,
|
||
"prompt": None,
|
||
"schedule": None,
|
||
"positionals": [],
|
||
}
|
||
i = 0
|
||
while i < len(tokens):
|
||
token = tokens[i]
|
||
if token == "--name" and i + 1 < len(tokens):
|
||
opts["name"] = tokens[i + 1]
|
||
i += 2
|
||
elif token == "--deliver" and i + 1 < len(tokens):
|
||
opts["deliver"] = tokens[i + 1]
|
||
i += 2
|
||
elif token == "--repeat" and i + 1 < len(tokens):
|
||
try:
|
||
opts["repeat"] = int(tokens[i + 1])
|
||
except ValueError:
|
||
print("(._.) --repeat must be an integer")
|
||
return None
|
||
i += 2
|
||
elif token == "--skill" and i + 1 < len(tokens):
|
||
opts["skills"].append(tokens[i + 1])
|
||
i += 2
|
||
elif token == "--add-skill" and i + 1 < len(tokens):
|
||
opts["add_skills"].append(tokens[i + 1])
|
||
i += 2
|
||
elif token == "--remove-skill" and i + 1 < len(tokens):
|
||
opts["remove_skills"].append(tokens[i + 1])
|
||
i += 2
|
||
elif token == "--clear-skills":
|
||
opts["clear_skills"] = True
|
||
i += 1
|
||
elif token == "--all":
|
||
opts["all"] = True
|
||
i += 1
|
||
elif token == "--prompt" and i + 1 < len(tokens):
|
||
opts["prompt"] = tokens[i + 1]
|
||
i += 2
|
||
elif token == "--schedule" and i + 1 < len(tokens):
|
||
opts["schedule"] = tokens[i + 1]
|
||
i += 2
|
||
else:
|
||
opts["positionals"].append(token)
|
||
i += 1
|
||
return opts
|
||
|
||
tokens = shlex.split(cmd)
|
||
|
||
if len(tokens) == 1:
|
||
print()
|
||
print("+" + "-" * 68 + "+")
|
||
print("|" + " " * 22 + "(^_^) Scheduled Tasks" + " " * 23 + "|")
|
||
print("+" + "-" * 68 + "+")
|
||
print()
|
||
print(" Commands:")
|
||
print(" /cron list")
|
||
print(' /cron add "every 2h" "Check server status" [--skill blogwatcher]')
|
||
print(' /cron edit <job_id> --schedule "every 4h" --prompt "New task"')
|
||
print(" /cron edit <job_id> --skill blogwatcher --skill find-nearby")
|
||
print(" /cron edit <job_id> --remove-skill blogwatcher")
|
||
print(" /cron edit <job_id> --clear-skills")
|
||
print(" /cron pause <job_id>")
|
||
print(" /cron resume <job_id>")
|
||
print(" /cron run <job_id>")
|
||
print(" /cron remove <job_id>")
|
||
print()
|
||
result = _cron_api(action="list")
|
||
jobs = result.get("jobs", []) if result.get("success") else []
|
||
if jobs:
|
||
print(" Current Jobs:")
|
||
print(" " + "-" * 63)
|
||
for job in jobs:
|
||
repeat_str = job.get("repeat", "?")
|
||
print(f" {job['job_id'][:12]:<12} | {job['schedule']:<15} | {repeat_str:<8}")
|
||
if job.get("skills"):
|
||
print(f" Skills: {', '.join(job['skills'])}")
|
||
print(f" {job.get('prompt_preview', '')}")
|
||
if job.get("next_run_at"):
|
||
print(f" Next: {job['next_run_at']}")
|
||
print()
|
||
else:
|
||
print(" No scheduled jobs. Use '/cron add' to create one.")
|
||
print()
|
||
return
|
||
|
||
subcommand = tokens[1].lower()
|
||
opts = _parse_flags(tokens[2:])
|
||
if opts is None:
|
||
return
|
||
|
||
if subcommand == "list":
|
||
result = _cron_api(action="list", include_disabled=opts["all"])
|
||
jobs = result.get("jobs", []) if result.get("success") else []
|
||
if not jobs:
|
||
print("(._.) No scheduled jobs.")
|
||
return
|
||
|
||
print()
|
||
print("Scheduled Jobs:")
|
||
print("-" * 80)
|
||
for job in jobs:
|
||
print(f" ID: {job['job_id']}")
|
||
print(f" Name: {job['name']}")
|
||
print(f" State: {job.get('state', '?')}")
|
||
print(f" Schedule: {job['schedule']} ({job.get('repeat', '?')})")
|
||
print(f" Next run: {job.get('next_run_at', 'N/A')}")
|
||
if job.get("skills"):
|
||
print(f" Skills: {', '.join(job['skills'])}")
|
||
print(f" Prompt: {job.get('prompt_preview', '')}")
|
||
if job.get("last_run_at"):
|
||
print(f" Last run: {job['last_run_at']} ({job.get('last_status', '?')})")
|
||
print()
|
||
return
|
||
|
||
if subcommand in {"add", "create"}:
|
||
positionals = opts["positionals"]
|
||
if not positionals:
|
||
print("(._.) Usage: /cron add <schedule> <prompt>")
|
||
return
|
||
schedule = opts["schedule"] or positionals[0]
|
||
prompt = opts["prompt"] or " ".join(positionals[1:])
|
||
skills = _normalize_skills(opts["skills"])
|
||
if not prompt and not skills:
|
||
print("(._.) Please provide a prompt or at least one skill")
|
||
return
|
||
result = _cron_api(
|
||
action="create",
|
||
schedule=schedule,
|
||
prompt=prompt or None,
|
||
name=opts["name"],
|
||
deliver=opts["deliver"],
|
||
repeat=opts["repeat"],
|
||
skills=skills or None,
|
||
)
|
||
if result.get("success"):
|
||
print(f"(^_^)b Created job: {result['job_id']}")
|
||
print(f" Schedule: {result['schedule']}")
|
||
if result.get("skills"):
|
||
print(f" Skills: {', '.join(result['skills'])}")
|
||
print(f" Next run: {result['next_run_at']}")
|
||
else:
|
||
print(f"(x_x) Failed to create job: {result.get('error')}")
|
||
return
|
||
|
||
if subcommand == "edit":
|
||
positionals = opts["positionals"]
|
||
if not positionals:
|
||
print("(._.) Usage: /cron edit <job_id> [--schedule ...] [--prompt ...] [--skill ...]")
|
||
return
|
||
job_id = positionals[0]
|
||
existing = get_job(job_id)
|
||
if not existing:
|
||
print(f"(._.) Job not found: {job_id}")
|
||
return
|
||
|
||
final_skills = None
|
||
replacement_skills = _normalize_skills(opts["skills"])
|
||
add_skills = _normalize_skills(opts["add_skills"])
|
||
remove_skills = set(_normalize_skills(opts["remove_skills"]))
|
||
existing_skills = list(existing.get("skills") or ([] if not existing.get("skill") else [existing.get("skill")]))
|
||
if opts["clear_skills"]:
|
||
final_skills = []
|
||
elif replacement_skills:
|
||
final_skills = replacement_skills
|
||
elif add_skills or remove_skills:
|
||
final_skills = [skill for skill in existing_skills if skill not in remove_skills]
|
||
for skill in add_skills:
|
||
if skill not in final_skills:
|
||
final_skills.append(skill)
|
||
|
||
result = _cron_api(
|
||
action="update",
|
||
job_id=job_id,
|
||
schedule=opts["schedule"],
|
||
prompt=opts["prompt"],
|
||
name=opts["name"],
|
||
deliver=opts["deliver"],
|
||
repeat=opts["repeat"],
|
||
skills=final_skills,
|
||
)
|
||
if result.get("success"):
|
||
job = result["job"]
|
||
print(f"(^_^)b Updated job: {job['job_id']}")
|
||
print(f" Schedule: {job['schedule']}")
|
||
if job.get("skills"):
|
||
print(f" Skills: {', '.join(job['skills'])}")
|
||
else:
|
||
print(" Skills: none")
|
||
else:
|
||
print(f"(x_x) Failed to update job: {result.get('error')}")
|
||
return
|
||
|
||
if subcommand in {"pause", "resume", "run", "remove", "rm", "delete"}:
|
||
positionals = opts["positionals"]
|
||
if not positionals:
|
||
print(f"(._.) Usage: /cron {subcommand} <job_id>")
|
||
return
|
||
job_id = positionals[0]
|
||
action = "remove" if subcommand in {"remove", "rm", "delete"} else subcommand
|
||
result = _cron_api(action=action, job_id=job_id, reason="paused from /cron" if action == "pause" else None)
|
||
if not result.get("success"):
|
||
print(f"(x_x) Failed to {action} job: {result.get('error')}")
|
||
return
|
||
if action == "pause":
|
||
print(f"(^_^)b Paused job: {result['job']['name']} ({job_id})")
|
||
elif action == "resume":
|
||
print(f"(^_^)b Resumed job: {result['job']['name']} ({job_id})")
|
||
print(f" Next run: {result['job'].get('next_run_at')}")
|
||
elif action == "run":
|
||
print(f"(^_^)b Triggered job: {result['job']['name']} ({job_id})")
|
||
print(" It will run on the next scheduler tick.")
|
||
else:
|
||
removed = result.get("removed_job", {})
|
||
print(f"(^_^)b Removed job: {removed.get('name', job_id)} ({job_id})")
|
||
return
|
||
|
||
print(f"(._.) Unknown cron command: {subcommand}")
|
||
print(" Available: list, add, edit, pause, resume, run, remove")
|
||
|
||
def _handle_skills_command(self, cmd: str):
|
||
"""Handle /skills slash command — delegates to hermes_cli.skills_hub."""
|
||
from hermes_cli.skills_hub import handle_skills_slash
|
||
handle_skills_slash(cmd, ChatConsole())
|
||
|
||
def _show_gateway_status(self):
|
||
"""Show status of the gateway and connected messaging platforms."""
|
||
from gateway.config import load_gateway_config, Platform
|
||
|
||
print()
|
||
print("+" + "-" * 60 + "+")
|
||
print("|" + " " * 15 + "(✿◠‿◠) Gateway Status" + " " * 17 + "|")
|
||
print("+" + "-" * 60 + "+")
|
||
print()
|
||
|
||
try:
|
||
config = load_gateway_config()
|
||
connected = config.get_connected_platforms()
|
||
|
||
print(" Messaging Platform Configuration:")
|
||
print(" " + "-" * 55)
|
||
|
||
platform_status = {
|
||
Platform.TELEGRAM: ("Telegram", "TELEGRAM_BOT_TOKEN"),
|
||
Platform.DISCORD: ("Discord", "DISCORD_BOT_TOKEN"),
|
||
Platform.WHATSAPP: ("WhatsApp", "WHATSAPP_ENABLED"),
|
||
}
|
||
|
||
for platform, (name, env_var) in platform_status.items():
|
||
pconfig = config.platforms.get(platform)
|
||
if pconfig and pconfig.enabled:
|
||
home = config.get_home_channel(platform)
|
||
home_str = f" → {home.name}" if home else ""
|
||
print(f" ✓ {name:<12} Enabled{home_str}")
|
||
else:
|
||
print(f" ○ {name:<12} Not configured ({env_var})")
|
||
|
||
print()
|
||
print(" Session Reset Policy:")
|
||
print(" " + "-" * 55)
|
||
policy = config.default_reset_policy
|
||
print(f" Mode: {policy.mode}")
|
||
print(f" Daily reset at: {policy.at_hour}:00")
|
||
print(f" Idle timeout: {policy.idle_minutes} minutes")
|
||
|
||
print()
|
||
print(" To start the gateway:")
|
||
print(" python cli.py --gateway")
|
||
print()
|
||
print(" Configuration file: ~/.hermes/config.yaml")
|
||
print()
|
||
|
||
except Exception as e:
|
||
print(f" Error loading gateway config: {e}")
|
||
print()
|
||
print(" To configure the gateway:")
|
||
print(" 1. Set environment variables:")
|
||
print(" TELEGRAM_BOT_TOKEN=your_token")
|
||
print(" DISCORD_BOT_TOKEN=your_token")
|
||
print(" 2. Or configure settings in ~/.hermes/config.yaml")
|
||
print()
|
||
|
||
def process_command(self, command: str) -> bool:
|
||
"""
|
||
Process a slash command.
|
||
|
||
Args:
|
||
command: The command string (starting with /)
|
||
|
||
Returns:
|
||
bool: True to continue, False to exit
|
||
"""
|
||
# Lowercase only for dispatch matching; preserve original case for arguments
|
||
cmd_lower = command.lower().strip()
|
||
cmd_original = command.strip()
|
||
|
||
# Resolve aliases via central registry so adding an alias is a one-line
|
||
# change in hermes_cli/commands.py instead of touching every dispatch site.
|
||
from hermes_cli.commands import resolve_command as _resolve_cmd
|
||
_base_word = cmd_lower.split()[0].lstrip("/")
|
||
_cmd_def = _resolve_cmd(_base_word)
|
||
canonical = _cmd_def.name if _cmd_def else _base_word
|
||
|
||
if canonical in ("quit", "exit", "q"):
|
||
return False
|
||
elif canonical == "help":
|
||
self.show_help()
|
||
elif canonical == "tools":
|
||
self._handle_tools_command(cmd_original)
|
||
elif canonical == "toolsets":
|
||
self.show_toolsets()
|
||
elif canonical == "config":
|
||
self.show_config()
|
||
elif canonical == "clear":
|
||
self.new_session(silent=True)
|
||
# Clear terminal screen. Inside the TUI, Rich's console.clear()
|
||
# goes through patch_stdout's StdoutProxy which swallows the
|
||
# screen-clear escape sequences. Use prompt_toolkit's output
|
||
# object directly to actually clear the terminal.
|
||
if self._app:
|
||
out = self._app.output
|
||
out.erase_screen()
|
||
out.cursor_goto(0, 0)
|
||
out.flush()
|
||
else:
|
||
self.console.clear()
|
||
# Show fresh banner. Inside the TUI we must route Rich output
|
||
# through ChatConsole (which uses prompt_toolkit's native ANSI
|
||
# renderer) instead of self.console (which writes raw to stdout
|
||
# and gets mangled by patch_stdout).
|
||
if self._app:
|
||
cc = ChatConsole()
|
||
term_w = shutil.get_terminal_size().columns
|
||
if self.compact or term_w < 80:
|
||
cc.print(_build_compact_banner())
|
||
else:
|
||
tools = get_tool_definitions(enabled_toolsets=self.enabled_toolsets, quiet_mode=True)
|
||
cwd = os.getenv("TERMINAL_CWD", os.getcwd())
|
||
ctx_len = None
|
||
if hasattr(self, 'agent') and self.agent and hasattr(self.agent, 'context_compressor'):
|
||
ctx_len = self.agent.context_compressor.context_length
|
||
build_welcome_banner(
|
||
console=cc,
|
||
model=self.model,
|
||
cwd=cwd,
|
||
tools=tools,
|
||
enabled_toolsets=self.enabled_toolsets,
|
||
session_id=self.session_id,
|
||
context_length=ctx_len,
|
||
)
|
||
_cprint(" ✨ (◕‿◕)✨ Fresh start! Screen cleared and conversation reset.\n")
|
||
else:
|
||
self.show_banner()
|
||
print(" ✨ (◕‿◕)✨ Fresh start! Screen cleared and conversation reset.\n")
|
||
elif canonical == "history":
|
||
self.show_history()
|
||
elif canonical == "title":
|
||
parts = cmd_original.split(maxsplit=1)
|
||
if len(parts) > 1:
|
||
raw_title = parts[1].strip()
|
||
if raw_title:
|
||
if self._session_db:
|
||
# Sanitize the title early so feedback matches what gets stored
|
||
try:
|
||
from hermes_state import SessionDB
|
||
new_title = SessionDB.sanitize_title(raw_title)
|
||
except ValueError as e:
|
||
_cprint(f" {e}")
|
||
new_title = None
|
||
if not new_title:
|
||
_cprint(" Title is empty after cleanup. Please use printable characters.")
|
||
elif self._session_db.get_session(self.session_id):
|
||
# Session exists in DB — set title directly
|
||
try:
|
||
if self._session_db.set_session_title(self.session_id, new_title):
|
||
_cprint(f" Session title set: {new_title}")
|
||
# Re-map Honcho session key to new title
|
||
if self.agent and getattr(self.agent, '_honcho', None):
|
||
try:
|
||
hcfg = self.agent._honcho_config
|
||
new_key = (
|
||
hcfg.resolve_session_name(
|
||
session_title=new_title,
|
||
session_id=self.agent.session_id,
|
||
)
|
||
if hcfg else new_title
|
||
)
|
||
if new_key and new_key != self.agent._honcho_session_key:
|
||
old_key = self.agent._honcho_session_key
|
||
self.agent._honcho.get_or_create(new_key)
|
||
self.agent._honcho_session_key = new_key
|
||
from tools.honcho_tools import set_session_context
|
||
set_session_context(self.agent._honcho, new_key)
|
||
from agent.display import honcho_session_line, write_tty
|
||
write_tty(honcho_session_line(hcfg.workspace_id, new_key) + "\n")
|
||
_cprint(f" Honcho session: {old_key} → {new_key}")
|
||
except Exception:
|
||
pass
|
||
else:
|
||
_cprint(" Session not found in database.")
|
||
except ValueError as e:
|
||
_cprint(f" {e}")
|
||
else:
|
||
# Session not created yet — defer the title
|
||
# Check uniqueness proactively with the sanitized title
|
||
existing = self._session_db.get_session_by_title(new_title)
|
||
if existing:
|
||
_cprint(f" Title '{new_title}' is already in use by session {existing['id']}")
|
||
else:
|
||
self._pending_title = new_title
|
||
_cprint(f" Session title queued: {new_title} (will be saved on first message)")
|
||
else:
|
||
_cprint(" Session database not available.")
|
||
else:
|
||
_cprint(" Usage: /title <your session title>")
|
||
else:
|
||
# Show current title and session ID if no argument given
|
||
if self._session_db:
|
||
_cprint(f" Session ID: {self.session_id}")
|
||
session = self._session_db.get_session(self.session_id)
|
||
if session and session.get("title"):
|
||
_cprint(f" Title: {session['title']}")
|
||
elif self._pending_title:
|
||
_cprint(f" Title (pending): {self._pending_title}")
|
||
else:
|
||
_cprint(f" No title set. Usage: /title <your session title>")
|
||
else:
|
||
_cprint(" Session database not available.")
|
||
elif canonical == "new":
|
||
self.new_session()
|
||
elif canonical == "model":
|
||
# Use original case so model names like "Anthropic/Claude-Opus-4" are preserved
|
||
parts = cmd_original.split(maxsplit=1)
|
||
if len(parts) > 1:
|
||
from hermes_cli.auth import resolve_provider
|
||
from hermes_cli.models import (
|
||
parse_model_input,
|
||
validate_requested_model,
|
||
_PROVIDER_LABELS,
|
||
)
|
||
|
||
raw_input = parts[1].strip()
|
||
|
||
# Parse provider:model syntax (e.g. "openrouter:anthropic/claude-sonnet-4.5")
|
||
current_provider = self.provider or self.requested_provider or "openrouter"
|
||
target_provider, new_model = parse_model_input(raw_input, current_provider)
|
||
# Auto-detect provider when no explicit provider:model syntax was used.
|
||
# Skip auto-detection for custom providers — the model name might
|
||
# coincidentally match a known provider's catalog, but the user
|
||
# intends to use it on their custom endpoint. Require explicit
|
||
# provider:model syntax (e.g. /model openai-codex:gpt-5.2-codex)
|
||
# to switch away from a custom endpoint.
|
||
_base = self.base_url or ""
|
||
is_custom = current_provider == "custom" or (
|
||
"localhost" in _base or "127.0.0.1" in _base
|
||
)
|
||
if target_provider == current_provider and not is_custom:
|
||
from hermes_cli.models import detect_provider_for_model
|
||
detected = detect_provider_for_model(new_model, current_provider)
|
||
if detected:
|
||
target_provider, new_model = detected
|
||
provider_changed = target_provider != current_provider
|
||
|
||
# If provider is changing, re-resolve credentials for the new provider
|
||
api_key_for_probe = self.api_key
|
||
base_url_for_probe = self.base_url
|
||
if provider_changed:
|
||
try:
|
||
from hermes_cli.runtime_provider import resolve_runtime_provider
|
||
runtime = resolve_runtime_provider(requested=target_provider)
|
||
api_key_for_probe = runtime.get("api_key", "")
|
||
base_url_for_probe = runtime.get("base_url", "")
|
||
except Exception as e:
|
||
provider_label = _PROVIDER_LABELS.get(target_provider, target_provider)
|
||
if target_provider == "custom":
|
||
print(f"(>_<) Custom endpoint not configured. Set OPENAI_BASE_URL and OPENAI_API_KEY,")
|
||
print(f" or run: hermes setup → Custom OpenAI-compatible endpoint")
|
||
else:
|
||
print(f"(>_<) Could not resolve credentials for provider '{provider_label}': {e}")
|
||
print(f"(^_^) Current model unchanged: {self.model}")
|
||
return True
|
||
|
||
try:
|
||
validation = validate_requested_model(
|
||
new_model,
|
||
target_provider,
|
||
api_key=api_key_for_probe,
|
||
base_url=base_url_for_probe,
|
||
)
|
||
except Exception:
|
||
validation = {"accepted": True, "persist": True, "recognized": False, "message": None}
|
||
|
||
if not validation.get("accepted"):
|
||
print(f"(>_<) {validation.get('message')}")
|
||
print(f" Model unchanged: {self.model}")
|
||
if "Did you mean" not in (validation.get("message") or ""):
|
||
print(" Tip: Use /model to see available models, /provider to see providers")
|
||
else:
|
||
self.model = new_model
|
||
self.agent = None # Force re-init
|
||
|
||
if provider_changed:
|
||
self.requested_provider = target_provider
|
||
self.provider = target_provider
|
||
self.api_key = api_key_for_probe
|
||
self.base_url = base_url_for_probe
|
||
|
||
provider_label = _PROVIDER_LABELS.get(target_provider, target_provider)
|
||
provider_note = f" [provider: {provider_label}]" if provider_changed else ""
|
||
|
||
if validation.get("persist"):
|
||
saved_model = save_config_value("model.default", new_model)
|
||
if provider_changed:
|
||
save_config_value("model.provider", target_provider)
|
||
if saved_model:
|
||
print(f"(^_^)b Model changed to: {new_model}{provider_note} (saved to config)")
|
||
else:
|
||
print(f"(^_^) Model changed to: {new_model}{provider_note} (this session only)")
|
||
else:
|
||
message = validation.get("message") or ""
|
||
print(f"(^_^) Model changed to: {new_model}{provider_note} (this session only)")
|
||
if message:
|
||
print(f" Reason: {message}")
|
||
print(" Note: Model will revert on restart. Use a verified model to save to config.")
|
||
|
||
# Helpful hint when staying on a custom endpoint
|
||
if is_custom and not provider_changed:
|
||
endpoint = self.base_url or "custom endpoint"
|
||
print(f" Endpoint: {endpoint}")
|
||
print(f" Tip: To switch providers, use /model provider:model")
|
||
print(f" e.g. /model openai-codex:gpt-5.2-codex")
|
||
else:
|
||
self._show_model_and_providers()
|
||
elif canonical == "provider":
|
||
self._show_model_and_providers()
|
||
elif canonical == "prompt":
|
||
# Use original case so prompt text isn't lowercased
|
||
self._handle_prompt_command(cmd_original)
|
||
elif canonical == "personality":
|
||
# Use original case (handler lowercases the personality name itself)
|
||
self._handle_personality_command(cmd_original)
|
||
elif canonical == "plan":
|
||
self._handle_plan_command(cmd_original)
|
||
elif canonical == "retry":
|
||
retry_msg = self.retry_last()
|
||
if retry_msg and hasattr(self, '_pending_input'):
|
||
# Re-queue the message so process_loop sends it to the agent
|
||
self._pending_input.put(retry_msg)
|
||
elif canonical == "undo":
|
||
self.undo_last()
|
||
elif canonical == "save":
|
||
self.save_conversation()
|
||
elif canonical == "cron":
|
||
self._handle_cron_command(cmd_original)
|
||
elif canonical == "skills":
|
||
with self._busy_command(self._slow_command_status(cmd_original)):
|
||
self._handle_skills_command(cmd_original)
|
||
elif canonical == "platforms":
|
||
self._show_gateway_status()
|
||
elif canonical == "statusbar":
|
||
self._status_bar_visible = not self._status_bar_visible
|
||
state = "visible" if self._status_bar_visible else "hidden"
|
||
self.console.print(f" Status bar {state}")
|
||
elif canonical == "verbose":
|
||
self._toggle_verbose()
|
||
elif canonical == "reasoning":
|
||
self._handle_reasoning_command(cmd_original)
|
||
elif canonical == "compress":
|
||
self._manual_compress()
|
||
elif canonical == "usage":
|
||
self._show_usage()
|
||
elif canonical == "insights":
|
||
self._show_insights(cmd_original)
|
||
elif canonical == "paste":
|
||
self._handle_paste_command()
|
||
elif canonical == "reload-mcp":
|
||
with self._busy_command(self._slow_command_status(cmd_original)):
|
||
self._reload_mcp()
|
||
elif canonical == "browser":
|
||
self._handle_browser_command(cmd_original)
|
||
elif canonical == "plugins":
|
||
try:
|
||
from hermes_cli.plugins import get_plugin_manager
|
||
mgr = get_plugin_manager()
|
||
plugins = mgr.list_plugins()
|
||
if not plugins:
|
||
print("No plugins installed.")
|
||
print(f"Drop plugin directories into ~/.hermes/plugins/ to get started.")
|
||
else:
|
||
print(f"Plugins ({len(plugins)}):")
|
||
for p in plugins:
|
||
status = "✓" if p["enabled"] else "✗"
|
||
version = f" v{p['version']}" if p["version"] else ""
|
||
tools = f"{p['tools']} tools" if p["tools"] else ""
|
||
hooks = f"{p['hooks']} hooks" if p["hooks"] else ""
|
||
parts = [x for x in [tools, hooks] if x]
|
||
detail = f" ({', '.join(parts)})" if parts else ""
|
||
error = f" — {p['error']}" if p["error"] else ""
|
||
print(f" {status} {p['name']}{version}{detail}{error}")
|
||
except Exception as e:
|
||
print(f"Plugin system error: {e}")
|
||
elif canonical == "rollback":
|
||
self._handle_rollback_command(cmd_original)
|
||
elif canonical == "stop":
|
||
self._handle_stop_command()
|
||
elif canonical == "background":
|
||
self._handle_background_command(cmd_original)
|
||
elif canonical == "queue":
|
||
if not self._agent_running:
|
||
_cprint(" /queue only works while Hermes is busy. Just type your message normally.")
|
||
else:
|
||
# Extract prompt after "/queue " or "/q "
|
||
parts = cmd_original.split(None, 1)
|
||
payload = parts[1].strip() if len(parts) > 1 else ""
|
||
if not payload:
|
||
_cprint(" Usage: /queue <prompt>")
|
||
else:
|
||
self._pending_input.put(payload)
|
||
_cprint(f" Queued for the next turn: {payload[:80]}{'...' if len(payload) > 80 else ''}")
|
||
elif canonical == "skin":
|
||
self._handle_skin_command(cmd_original)
|
||
elif canonical == "voice":
|
||
self._handle_voice_command(cmd_original)
|
||
else:
|
||
# Check for user-defined quick commands (bypass agent loop, no LLM call)
|
||
base_cmd = cmd_lower.split()[0]
|
||
quick_commands = self.config.get("quick_commands", {})
|
||
if base_cmd.lstrip("/") in quick_commands:
|
||
qcmd = quick_commands[base_cmd.lstrip("/")]
|
||
if qcmd.get("type") == "exec":
|
||
import subprocess
|
||
exec_cmd = qcmd.get("command", "")
|
||
if exec_cmd:
|
||
try:
|
||
result = subprocess.run(
|
||
exec_cmd, shell=True, capture_output=True,
|
||
text=True, timeout=30
|
||
)
|
||
output = result.stdout.strip() or result.stderr.strip()
|
||
if output:
|
||
self.console.print(_rich_text_from_ansi(output))
|
||
else:
|
||
self.console.print("[dim]Command returned no output[/]")
|
||
except subprocess.TimeoutExpired:
|
||
self.console.print("[bold red]Quick command timed out (30s)[/]")
|
||
except Exception as e:
|
||
self.console.print(f"[bold red]Quick command error: {e}[/]")
|
||
else:
|
||
self.console.print(f"[bold red]Quick command '{base_cmd}' has no command defined[/]")
|
||
elif qcmd.get("type") == "alias":
|
||
target = qcmd.get("target", "").strip()
|
||
if target:
|
||
target = target if target.startswith("/") else f"/{target}"
|
||
user_args = cmd_original[len(base_cmd):].strip()
|
||
aliased_command = f"{target} {user_args}".strip()
|
||
return self.process_command(aliased_command)
|
||
else:
|
||
self.console.print(f"[bold red]Quick command '{base_cmd}' has no target defined[/]")
|
||
else:
|
||
self.console.print(f"[bold red]Quick command '{base_cmd}' has unsupported type (supported: 'exec', 'alias')[/]")
|
||
# Check for skill slash commands (/gif-search, /axolotl, etc.)
|
||
elif base_cmd in _skill_commands:
|
||
user_instruction = cmd_original[len(base_cmd):].strip()
|
||
msg = build_skill_invocation_message(
|
||
base_cmd, user_instruction, task_id=self.session_id
|
||
)
|
||
if msg:
|
||
skill_name = _skill_commands[base_cmd]["name"]
|
||
print(f"\n⚡ Loading skill: {skill_name}")
|
||
if hasattr(self, '_pending_input'):
|
||
self._pending_input.put(msg)
|
||
else:
|
||
self.console.print(f"[bold red]Failed to load skill for {base_cmd}[/]")
|
||
else:
|
||
# Prefix matching: if input uniquely identifies one command, execute it.
|
||
# Matches against both built-in COMMANDS and installed skill commands so
|
||
# that execution-time resolution agrees with tab-completion.
|
||
from hermes_cli.commands import COMMANDS
|
||
typed_base = cmd_lower.split()[0]
|
||
all_known = set(COMMANDS) | set(_skill_commands)
|
||
matches = [c for c in all_known if c.startswith(typed_base)]
|
||
if len(matches) > 1:
|
||
# Prefer an exact match (typed the full command name)
|
||
exact = [c for c in matches if c == typed_base]
|
||
if len(exact) == 1:
|
||
matches = exact
|
||
else:
|
||
# Prefer the unique shortest match:
|
||
# /qui → /quit (5) wins over /quint-pipeline (15)
|
||
min_len = min(len(c) for c in matches)
|
||
shortest = [c for c in matches if len(c) == min_len]
|
||
if len(shortest) == 1:
|
||
matches = shortest
|
||
if len(matches) == 1:
|
||
# Expand the prefix to the full command name, preserving arguments.
|
||
# Guard against redispatching the same token to avoid infinite
|
||
# recursion when the expanded name still doesn't hit an exact branch
|
||
# (e.g. /config with extra args that are not yet handled above).
|
||
full_name = matches[0]
|
||
if full_name == typed_base:
|
||
# Already an exact token — no expansion possible; fall through
|
||
_cprint(f"\033[1;31mUnknown command: {cmd_lower}{_RST}")
|
||
_cprint(f"{_DIM}{_GOLD}Type /help for available commands{_RST}")
|
||
else:
|
||
remainder = cmd_original.strip()[len(typed_base):]
|
||
full_cmd = full_name + remainder
|
||
return self.process_command(full_cmd)
|
||
elif len(matches) > 1:
|
||
_cprint(f"{_GOLD}Ambiguous command: {cmd_lower}{_RST}")
|
||
_cprint(f"{_DIM}Did you mean: {', '.join(sorted(matches))}?{_RST}")
|
||
else:
|
||
_cprint(f"\033[1;31mUnknown command: {cmd_lower}{_RST}")
|
||
_cprint(f"{_DIM}{_GOLD}Type /help for available commands{_RST}")
|
||
|
||
return True
|
||
|
||
def _handle_plan_command(self, cmd: str):
|
||
"""Handle /plan [request] — load the bundled plan skill."""
|
||
parts = cmd.strip().split(maxsplit=1)
|
||
user_instruction = parts[1].strip() if len(parts) > 1 else ""
|
||
|
||
plan_path = build_plan_path(user_instruction)
|
||
msg = build_skill_invocation_message(
|
||
"/plan",
|
||
user_instruction,
|
||
task_id=self.session_id,
|
||
runtime_note=(
|
||
"Save the markdown plan with write_file to this exact relative path "
|
||
f"inside the active workspace/backend cwd: {plan_path}"
|
||
),
|
||
)
|
||
|
||
if not msg:
|
||
self.console.print("[bold red]Failed to load the bundled /plan skill[/]")
|
||
return
|
||
|
||
_cprint(f" 📝 Plan mode queued via skill. Markdown plan target: {plan_path}")
|
||
if hasattr(self, '_pending_input'):
|
||
self._pending_input.put(msg)
|
||
else:
|
||
self.console.print("[bold red]Plan mode unavailable: input queue not initialized[/]")
|
||
|
||
def _handle_background_command(self, cmd: str):
|
||
"""Handle /background <prompt> — run a prompt in a separate background session.
|
||
|
||
Spawns a new AIAgent in a background thread with its own session.
|
||
When it completes, prints the result to the CLI without modifying
|
||
the active session's conversation history.
|
||
"""
|
||
parts = cmd.strip().split(maxsplit=1)
|
||
if len(parts) < 2 or not parts[1].strip():
|
||
_cprint(" Usage: /background <prompt>")
|
||
_cprint(" Example: /background Summarize the top HN stories today")
|
||
_cprint(" The task runs in a separate session and results display here when done.")
|
||
return
|
||
|
||
prompt = parts[1].strip()
|
||
self._background_task_counter += 1
|
||
task_num = self._background_task_counter
|
||
task_id = f"bg_{datetime.now().strftime('%H%M%S')}_{uuid.uuid4().hex[:6]}"
|
||
|
||
# Make sure we have valid credentials
|
||
if not self._ensure_runtime_credentials():
|
||
_cprint(" (>_<) Cannot start background task: no valid credentials.")
|
||
return
|
||
|
||
_cprint(f" 🔄 Background task #{task_num} started: \"{prompt[:60]}{'...' if len(prompt) > 60 else ''}\"")
|
||
_cprint(f" Task ID: {task_id}")
|
||
_cprint(f" You can continue chatting — results will appear when done.\n")
|
||
|
||
turn_route = self._resolve_turn_agent_config(prompt)
|
||
|
||
def run_background():
|
||
try:
|
||
bg_agent = AIAgent(
|
||
model=turn_route["model"],
|
||
api_key=turn_route["runtime"].get("api_key"),
|
||
base_url=turn_route["runtime"].get("base_url"),
|
||
provider=turn_route["runtime"].get("provider"),
|
||
api_mode=turn_route["runtime"].get("api_mode"),
|
||
acp_command=turn_route["runtime"].get("command"),
|
||
acp_args=turn_route["runtime"].get("args"),
|
||
max_iterations=self.max_turns,
|
||
enabled_toolsets=self.enabled_toolsets,
|
||
quiet_mode=True,
|
||
verbose_logging=False,
|
||
session_id=task_id,
|
||
platform="cli",
|
||
session_db=self._session_db,
|
||
reasoning_config=self.reasoning_config,
|
||
providers_allowed=self._providers_only,
|
||
providers_ignored=self._providers_ignore,
|
||
providers_order=self._providers_order,
|
||
provider_sort=self._provider_sort,
|
||
provider_require_parameters=self._provider_require_params,
|
||
provider_data_collection=self._provider_data_collection,
|
||
fallback_model=self._fallback_model,
|
||
)
|
||
|
||
result = bg_agent.run_conversation(
|
||
user_message=prompt,
|
||
task_id=task_id,
|
||
)
|
||
|
||
response = result.get("final_response", "") if result else ""
|
||
if not response and result and result.get("error"):
|
||
response = f"Error: {result['error']}"
|
||
|
||
# Display result in the CLI (thread-safe via patch_stdout)
|
||
print()
|
||
ChatConsole().print(f"[{_accent_hex()}]{'─' * 40}[/]")
|
||
_cprint(f" ✅ Background task #{task_num} complete")
|
||
_cprint(f" Prompt: \"{prompt[:60]}{'...' if len(prompt) > 60 else ''}\"")
|
||
ChatConsole().print(f"[{_accent_hex()}]{'─' * 40}[/]")
|
||
if response:
|
||
try:
|
||
from hermes_cli.skin_engine import get_active_skin
|
||
_skin = get_active_skin()
|
||
label = _skin.get_branding("response_label", "⚕ Hermes")
|
||
_resp_color = _skin.get_color("response_border", "#CD7F32")
|
||
_resp_text = _skin.get_color("banner_text", "#FFF8DC")
|
||
except Exception:
|
||
label = "⚕ Hermes"
|
||
_resp_color = "#CD7F32"
|
||
_resp_text = "#FFF8DC"
|
||
|
||
_chat_console = ChatConsole()
|
||
_chat_console.print(Panel(
|
||
_rich_text_from_ansi(response),
|
||
title=f"[{_resp_color} bold]{label} (background #{task_num})[/]",
|
||
title_align="left",
|
||
border_style=_resp_color,
|
||
style=_resp_text,
|
||
box=rich_box.HORIZONTALS,
|
||
padding=(1, 2),
|
||
))
|
||
else:
|
||
_cprint(" (No response generated)")
|
||
|
||
# Play bell if enabled
|
||
if self.bell_on_complete:
|
||
sys.stdout.write("\a")
|
||
sys.stdout.flush()
|
||
|
||
except Exception as e:
|
||
print()
|
||
_cprint(f" ❌ Background task #{task_num} failed: {e}")
|
||
finally:
|
||
self._background_tasks.pop(task_id, None)
|
||
if self._app:
|
||
self._invalidate(min_interval=0)
|
||
|
||
thread = threading.Thread(target=run_background, daemon=True, name=f"bg-task-{task_id}")
|
||
self._background_tasks[task_id] = thread
|
||
thread.start()
|
||
|
||
@staticmethod
|
||
def _try_launch_chrome_debug(port: int, system: str) -> bool:
|
||
"""Try to launch Chrome/Chromium with remote debugging enabled.
|
||
|
||
Returns True if a launch command was executed (doesn't guarantee success).
|
||
"""
|
||
import shutil
|
||
import subprocess as _sp
|
||
|
||
candidates = []
|
||
if system == "Darwin":
|
||
# macOS: try common app bundle locations
|
||
for app in (
|
||
"/Applications/Google Chrome.app/Contents/MacOS/Google Chrome",
|
||
"/Applications/Chromium.app/Contents/MacOS/Chromium",
|
||
"/Applications/Brave Browser.app/Contents/MacOS/Brave Browser",
|
||
"/Applications/Microsoft Edge.app/Contents/MacOS/Microsoft Edge",
|
||
):
|
||
if os.path.isfile(app):
|
||
candidates.append(app)
|
||
else:
|
||
# Linux: try common binary names
|
||
for name in ("google-chrome", "google-chrome-stable", "chromium-browser",
|
||
"chromium", "brave-browser", "microsoft-edge"):
|
||
path = shutil.which(name)
|
||
if path:
|
||
candidates.append(path)
|
||
|
||
if not candidates:
|
||
return False
|
||
|
||
chrome = candidates[0]
|
||
try:
|
||
_sp.Popen(
|
||
[chrome, f"--remote-debugging-port={port}"],
|
||
stdout=_sp.DEVNULL,
|
||
stderr=_sp.DEVNULL,
|
||
start_new_session=True, # detach from terminal
|
||
)
|
||
return True
|
||
except Exception:
|
||
return False
|
||
|
||
def _handle_browser_command(self, cmd: str):
|
||
"""Handle /browser connect|disconnect|status — manage live Chrome CDP connection."""
|
||
import platform as _plat
|
||
import subprocess as _sp
|
||
|
||
parts = cmd.strip().split(None, 1)
|
||
sub = parts[1].lower().strip() if len(parts) > 1 else "status"
|
||
|
||
_DEFAULT_CDP = "http://localhost:9222"
|
||
current = os.environ.get("BROWSER_CDP_URL", "").strip()
|
||
|
||
if sub.startswith("connect"):
|
||
# Optionally accept a custom CDP URL: /browser connect ws://host:port
|
||
connect_parts = cmd.strip().split(None, 2) # ["/browser", "connect", "ws://..."]
|
||
cdp_url = connect_parts[2].strip() if len(connect_parts) > 2 else _DEFAULT_CDP
|
||
|
||
# Clear any existing browser sessions so the next tool call uses the new backend
|
||
try:
|
||
from tools.browser_tool import cleanup_all_browsers
|
||
cleanup_all_browsers()
|
||
except Exception:
|
||
pass
|
||
|
||
print()
|
||
|
||
# Extract port for connectivity checks
|
||
_port = 9222
|
||
try:
|
||
_port = int(cdp_url.rsplit(":", 1)[-1].split("/")[0])
|
||
except (ValueError, IndexError):
|
||
pass
|
||
|
||
# Check if Chrome is already listening on the debug port
|
||
import socket
|
||
_already_open = False
|
||
try:
|
||
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
|
||
s.settimeout(1)
|
||
s.connect(("127.0.0.1", _port))
|
||
s.close()
|
||
_already_open = True
|
||
except (OSError, socket.timeout):
|
||
pass
|
||
|
||
if _already_open:
|
||
print(f" ✓ Chrome is already listening on port {_port}")
|
||
elif cdp_url == _DEFAULT_CDP:
|
||
# Try to auto-launch Chrome with remote debugging
|
||
print(" Chrome isn't running with remote debugging — attempting to launch...")
|
||
_launched = self._try_launch_chrome_debug(_port, _plat.system())
|
||
if _launched:
|
||
# Wait for the port to come up
|
||
import time as _time
|
||
for _wait in range(10):
|
||
try:
|
||
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
|
||
s.settimeout(1)
|
||
s.connect(("127.0.0.1", _port))
|
||
s.close()
|
||
_already_open = True
|
||
break
|
||
except (OSError, socket.timeout):
|
||
_time.sleep(0.5)
|
||
if _already_open:
|
||
print(f" ✓ Chrome launched and listening on port {_port}")
|
||
else:
|
||
print(f" ⚠ Chrome launched but port {_port} isn't responding yet")
|
||
print(" You may need to close existing Chrome windows first and retry")
|
||
else:
|
||
print(f" ⚠ Could not auto-launch Chrome")
|
||
# Show manual instructions as fallback
|
||
sys_name = _plat.system()
|
||
if sys_name == "Darwin":
|
||
chrome_cmd = 'open -a "Google Chrome" --args --remote-debugging-port=9222'
|
||
elif sys_name == "Windows":
|
||
chrome_cmd = 'chrome.exe --remote-debugging-port=9222'
|
||
else:
|
||
chrome_cmd = "google-chrome --remote-debugging-port=9222"
|
||
print(f" Launch Chrome manually: {chrome_cmd}")
|
||
else:
|
||
print(f" ⚠ Port {_port} is not reachable at {cdp_url}")
|
||
|
||
os.environ["BROWSER_CDP_URL"] = cdp_url
|
||
print()
|
||
print("🌐 Browser connected to live Chrome via CDP")
|
||
print(f" Endpoint: {cdp_url}")
|
||
print()
|
||
|
||
# Inject context message so the model knows
|
||
if hasattr(self, '_pending_input'):
|
||
self._pending_input.put(
|
||
"[System note: The user has connected your browser tools to their live Chrome browser "
|
||
"via Chrome DevTools Protocol. Your browser_navigate, browser_snapshot, browser_click, "
|
||
"and other browser tools now control their real browser — including any pages they have "
|
||
"open, logged-in sessions, and cookies. They likely opened specific sites or logged into "
|
||
"services before connecting. Please await their instruction before attempting to operate "
|
||
"the browser. When you do act, be mindful that your actions affect their real browser — "
|
||
"don't close tabs or navigate away from pages without asking.]"
|
||
)
|
||
|
||
elif sub == "disconnect":
|
||
if current:
|
||
os.environ.pop("BROWSER_CDP_URL", None)
|
||
try:
|
||
from tools.browser_tool import cleanup_all_browsers
|
||
cleanup_all_browsers()
|
||
except Exception:
|
||
pass
|
||
print()
|
||
print("🌐 Browser disconnected from live Chrome")
|
||
print(" Browser tools reverted to default mode (local headless or Browserbase)")
|
||
print()
|
||
|
||
if hasattr(self, '_pending_input'):
|
||
self._pending_input.put(
|
||
"[System note: The user has disconnected the browser tools from their live Chrome. "
|
||
"Browser tools are back to default mode (headless local browser or Browserbase cloud).]"
|
||
)
|
||
else:
|
||
print()
|
||
print("Browser is not connected to live Chrome (already using default mode)")
|
||
print()
|
||
|
||
elif sub == "status":
|
||
print()
|
||
if current:
|
||
print(f"🌐 Browser: connected to live Chrome via CDP")
|
||
print(f" Endpoint: {current}")
|
||
|
||
_port = 9222
|
||
try:
|
||
_port = int(current.rsplit(":", 1)[-1].split("/")[0])
|
||
except (ValueError, IndexError):
|
||
pass
|
||
try:
|
||
import socket
|
||
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
|
||
s.settimeout(1)
|
||
s.connect(("127.0.0.1", _port))
|
||
s.close()
|
||
print(f" Status: ✓ reachable")
|
||
except (OSError, Exception):
|
||
print(f" Status: ⚠ not reachable (Chrome may not be running)")
|
||
elif os.environ.get("BROWSERBASE_API_KEY"):
|
||
print("🌐 Browser: Browserbase (cloud)")
|
||
else:
|
||
print("🌐 Browser: local headless Chromium (agent-browser)")
|
||
print()
|
||
print(" /browser connect — connect to your live Chrome")
|
||
print(" /browser disconnect — revert to default")
|
||
print()
|
||
|
||
else:
|
||
print()
|
||
print("Usage: /browser connect|disconnect|status")
|
||
print()
|
||
print(" connect Connect browser tools to your live Chrome session")
|
||
print(" disconnect Revert to default browser backend")
|
||
print(" status Show current browser mode")
|
||
print()
|
||
|
||
def _handle_skin_command(self, cmd: str):
|
||
"""Handle /skin [name] — show or change the display skin."""
|
||
try:
|
||
from hermes_cli.skin_engine import list_skins, set_active_skin, get_active_skin_name
|
||
except ImportError:
|
||
print("Skin engine not available.")
|
||
return
|
||
|
||
parts = cmd.strip().split(maxsplit=1)
|
||
if len(parts) < 2 or not parts[1].strip():
|
||
# Show current skin and list available
|
||
current = get_active_skin_name()
|
||
skins = list_skins()
|
||
print(f"\n Current skin: {current}")
|
||
print(f" Available skins:")
|
||
for s in skins:
|
||
marker = " ●" if s["name"] == current else " "
|
||
source = f" ({s['source']})" if s["source"] == "user" else ""
|
||
print(f" {marker} {s['name']}{source} — {s['description']}")
|
||
print(f"\n Usage: /skin <name>")
|
||
print(f" Custom skins: drop a YAML file in ~/.hermes/skins/\n")
|
||
return
|
||
|
||
new_skin = parts[1].strip().lower()
|
||
available = {s["name"] for s in list_skins()}
|
||
if new_skin not in available:
|
||
print(f" Unknown skin: {new_skin}")
|
||
print(f" Available: {', '.join(sorted(available))}")
|
||
return
|
||
|
||
set_active_skin(new_skin)
|
||
if save_config_value("display.skin", new_skin):
|
||
print(f" Skin set to: {new_skin} (saved)")
|
||
else:
|
||
print(f" Skin set to: {new_skin}")
|
||
print(" Note: banner colors will update on next session start.")
|
||
if self._apply_tui_skin_style():
|
||
print(" Prompt + TUI colors updated.")
|
||
|
||
def _toggle_verbose(self):
|
||
"""Cycle tool progress mode: off → new → all → verbose → off."""
|
||
cycle = ["off", "new", "all", "verbose"]
|
||
try:
|
||
idx = cycle.index(self.tool_progress_mode)
|
||
except ValueError:
|
||
idx = 2 # default to "all"
|
||
self.tool_progress_mode = cycle[(idx + 1) % len(cycle)]
|
||
self.verbose = self.tool_progress_mode == "verbose"
|
||
|
||
if self.agent:
|
||
self.agent.verbose_logging = self.verbose
|
||
self.agent.quiet_mode = not self.verbose
|
||
# Auto-enable reasoning display in verbose mode
|
||
if self.verbose:
|
||
self.agent.reasoning_callback = self._on_reasoning
|
||
elif not self.show_reasoning:
|
||
self.agent.reasoning_callback = None
|
||
|
||
labels = {
|
||
"off": "[dim]Tool progress: OFF[/] — silent mode, just the final response.",
|
||
"new": "[yellow]Tool progress: NEW[/] — show each new tool (skip repeats).",
|
||
"all": "[green]Tool progress: ALL[/] — show every tool call.",
|
||
"verbose": "[bold green]Tool progress: VERBOSE[/] — full args, results, think blocks, and debug logs.",
|
||
}
|
||
self.console.print(labels.get(self.tool_progress_mode, ""))
|
||
|
||
def _handle_reasoning_command(self, cmd: str):
|
||
"""Handle /reasoning — manage effort level and display toggle.
|
||
|
||
Usage:
|
||
/reasoning Show current effort level and display state
|
||
/reasoning <level> Set reasoning effort (none, low, medium, high, xhigh)
|
||
/reasoning show|on Show model thinking/reasoning in output
|
||
/reasoning hide|off Hide model thinking/reasoning from output
|
||
"""
|
||
parts = cmd.strip().split(maxsplit=1)
|
||
|
||
if len(parts) < 2:
|
||
# Show current state
|
||
rc = self.reasoning_config
|
||
if rc is None:
|
||
level = "medium (default)"
|
||
elif rc.get("enabled") is False:
|
||
level = "none (disabled)"
|
||
else:
|
||
level = rc.get("effort", "medium")
|
||
display_state = "on ✓" if self.show_reasoning else "off"
|
||
_cprint(f" {_GOLD}Reasoning effort: {level}{_RST}")
|
||
_cprint(f" {_GOLD}Reasoning display: {display_state}{_RST}")
|
||
_cprint(f" {_DIM}Usage: /reasoning <none|low|medium|high|xhigh|show|hide>{_RST}")
|
||
return
|
||
|
||
arg = parts[1].strip().lower()
|
||
|
||
# Display toggle
|
||
if arg in ("show", "on"):
|
||
self.show_reasoning = True
|
||
if self.agent:
|
||
self.agent.reasoning_callback = self._on_reasoning
|
||
save_config_value("display.show_reasoning", True)
|
||
_cprint(f" {_GOLD}✓ Reasoning display: ON (saved){_RST}")
|
||
_cprint(f" {_DIM} Model thinking will be shown during and after each response.{_RST}")
|
||
return
|
||
if arg in ("hide", "off"):
|
||
self.show_reasoning = False
|
||
if self.agent:
|
||
self.agent.reasoning_callback = None
|
||
save_config_value("display.show_reasoning", False)
|
||
_cprint(f" {_GOLD}✓ Reasoning display: OFF (saved){_RST}")
|
||
return
|
||
|
||
# Effort level change
|
||
parsed = _parse_reasoning_config(arg)
|
||
if parsed is None:
|
||
_cprint(f" {_DIM}(._.) Unknown argument: {arg}{_RST}")
|
||
_cprint(f" {_DIM}Valid levels: none, low, minimal, medium, high, xhigh{_RST}")
|
||
_cprint(f" {_DIM}Display: show, hide{_RST}")
|
||
return
|
||
|
||
self.reasoning_config = parsed
|
||
self.agent = None # Force agent re-init with new reasoning config
|
||
|
||
if save_config_value("agent.reasoning_effort", arg):
|
||
_cprint(f" {_GOLD}✓ Reasoning effort set to '{arg}' (saved to config){_RST}")
|
||
else:
|
||
_cprint(f" {_GOLD}✓ Reasoning effort set to '{arg}' (session only){_RST}")
|
||
|
||
def _on_reasoning(self, reasoning_text: str):
|
||
"""Callback for intermediate reasoning display during tool-call loops."""
|
||
if self.verbose:
|
||
# Verbose mode: show full reasoning text
|
||
_cprint(f" {_DIM}[thinking] {reasoning_text.strip()}{_RST}")
|
||
else:
|
||
lines = reasoning_text.strip().splitlines()
|
||
if len(lines) > 5:
|
||
preview = "\n".join(lines[:5])
|
||
preview += f"\n ... ({len(lines) - 5} more lines)"
|
||
else:
|
||
preview = reasoning_text.strip()
|
||
_cprint(f" {_DIM}[thinking] {preview}{_RST}")
|
||
|
||
def _manual_compress(self):
|
||
"""Manually trigger context compression on the current conversation."""
|
||
if not self.conversation_history or len(self.conversation_history) < 4:
|
||
print("(._.) Not enough conversation to compress (need at least 4 messages).")
|
||
return
|
||
|
||
if not self.agent:
|
||
print("(._.) No active agent -- send a message first.")
|
||
return
|
||
|
||
if not self.agent.compression_enabled:
|
||
print("(._.) Compression is disabled in config.")
|
||
return
|
||
|
||
original_count = len(self.conversation_history)
|
||
try:
|
||
from agent.model_metadata import estimate_messages_tokens_rough
|
||
approx_tokens = estimate_messages_tokens_rough(self.conversation_history)
|
||
print(f"🗜️ Compressing {original_count} messages (~{approx_tokens:,} tokens)...")
|
||
|
||
compressed, new_system = self.agent._compress_context(
|
||
self.conversation_history,
|
||
self.agent._cached_system_prompt or "",
|
||
approx_tokens=approx_tokens,
|
||
)
|
||
self.conversation_history = compressed
|
||
new_count = len(self.conversation_history)
|
||
new_tokens = estimate_messages_tokens_rough(self.conversation_history)
|
||
print(
|
||
f" ✅ Compressed: {original_count} → {new_count} messages "
|
||
f"(~{approx_tokens:,} → ~{new_tokens:,} tokens)"
|
||
)
|
||
# Flush Honcho async queue so queued messages land before context resets
|
||
if self.agent and getattr(self.agent, '_honcho', None):
|
||
try:
|
||
self.agent._honcho.flush_all()
|
||
except Exception:
|
||
pass
|
||
except Exception as e:
|
||
print(f" ❌ Compression failed: {e}")
|
||
|
||
def _show_usage(self):
|
||
"""Show cumulative token usage for the current session."""
|
||
if not self.agent:
|
||
print("(._.) No active agent -- send a message first.")
|
||
return
|
||
|
||
agent = self.agent
|
||
input_tokens = getattr(agent, "session_input_tokens", 0) or 0
|
||
output_tokens = getattr(agent, "session_output_tokens", 0) or 0
|
||
cache_read_tokens = getattr(agent, "session_cache_read_tokens", 0) or 0
|
||
cache_write_tokens = getattr(agent, "session_cache_write_tokens", 0) or 0
|
||
prompt = agent.session_prompt_tokens
|
||
completion = agent.session_completion_tokens
|
||
total = agent.session_total_tokens
|
||
calls = agent.session_api_calls
|
||
|
||
if calls == 0:
|
||
print("(._.) No API calls made yet in this session.")
|
||
return
|
||
|
||
# Current context window state
|
||
compressor = agent.context_compressor
|
||
last_prompt = compressor.last_prompt_tokens
|
||
ctx_len = compressor.context_length
|
||
pct = (last_prompt / ctx_len * 100) if ctx_len else 0
|
||
compressions = compressor.compression_count
|
||
|
||
msg_count = len(self.conversation_history)
|
||
cost_result = estimate_usage_cost(
|
||
agent.model,
|
||
CanonicalUsage(
|
||
input_tokens=input_tokens,
|
||
output_tokens=output_tokens,
|
||
cache_read_tokens=cache_read_tokens,
|
||
cache_write_tokens=cache_write_tokens,
|
||
),
|
||
provider=getattr(agent, "provider", None),
|
||
base_url=getattr(agent, "base_url", None),
|
||
)
|
||
elapsed = format_duration_compact((datetime.now() - self.session_start).total_seconds())
|
||
|
||
print(f" 📊 Session Token Usage")
|
||
print(f" {'─' * 40}")
|
||
print(f" Model: {agent.model}")
|
||
print(f" Input tokens: {input_tokens:>10,}")
|
||
print(f" Cache read tokens: {cache_read_tokens:>10,}")
|
||
print(f" Cache write tokens: {cache_write_tokens:>10,}")
|
||
print(f" Output tokens: {output_tokens:>10,}")
|
||
print(f" Prompt tokens (total): {prompt:>10,}")
|
||
print(f" Completion tokens: {completion:>10,}")
|
||
print(f" Total tokens: {total:>10,}")
|
||
print(f" API calls: {calls:>10,}")
|
||
print(f" Session duration: {elapsed:>10}")
|
||
print(f" Cost status: {cost_result.status:>10}")
|
||
print(f" Cost source: {cost_result.source:>10}")
|
||
if cost_result.amount_usd is not None:
|
||
prefix = "~" if cost_result.status == "estimated" else ""
|
||
print(f" Total cost: {prefix}${float(cost_result.amount_usd):>10.4f}")
|
||
elif cost_result.status == "included":
|
||
print(f" Total cost: {'included':>10}")
|
||
else:
|
||
print(f" Total cost: {'n/a':>10}")
|
||
print(f" {'─' * 40}")
|
||
print(f" Current context: {last_prompt:,} / {ctx_len:,} ({pct:.0f}%)")
|
||
print(f" Messages: {msg_count}")
|
||
print(f" Compressions: {compressions}")
|
||
if cost_result.status == "unknown":
|
||
print(f" Note: Pricing unknown for {agent.model}")
|
||
|
||
if self.verbose:
|
||
logging.getLogger().setLevel(logging.DEBUG)
|
||
for noisy in ('openai', 'openai._base_client', 'httpx', 'httpcore', 'asyncio', 'hpack', 'grpc', 'modal'):
|
||
logging.getLogger(noisy).setLevel(logging.WARNING)
|
||
else:
|
||
logging.getLogger().setLevel(logging.INFO)
|
||
for quiet_logger in ('tools', 'minisweagent', 'run_agent', 'trajectory_compressor', 'cron', 'hermes_cli'):
|
||
logging.getLogger(quiet_logger).setLevel(logging.ERROR)
|
||
|
||
def _show_insights(self, command: str = "/insights"):
|
||
"""Show usage insights and analytics from session history."""
|
||
# Parse optional --days flag
|
||
parts = command.split()
|
||
days = 30
|
||
source = None
|
||
i = 1
|
||
while i < len(parts):
|
||
if parts[i] == "--days" and i + 1 < len(parts):
|
||
try:
|
||
days = int(parts[i + 1])
|
||
except ValueError:
|
||
print(f" Invalid --days value: {parts[i + 1]}")
|
||
return
|
||
i += 2
|
||
elif parts[i] == "--source" and i + 1 < len(parts):
|
||
source = parts[i + 1]
|
||
i += 2
|
||
else:
|
||
i += 1
|
||
|
||
try:
|
||
from hermes_state import SessionDB
|
||
from agent.insights import InsightsEngine
|
||
|
||
db = SessionDB()
|
||
engine = InsightsEngine(db)
|
||
report = engine.generate(days=days, source=source)
|
||
print(engine.format_terminal(report))
|
||
db.close()
|
||
except Exception as e:
|
||
print(f" Error generating insights: {e}")
|
||
|
||
def _check_config_mcp_changes(self) -> None:
|
||
"""Detect mcp_servers changes in config.yaml and auto-reload MCP connections.
|
||
|
||
Called from process_loop every CONFIG_WATCH_INTERVAL seconds.
|
||
Compares config.yaml mtime + mcp_servers section against the last
|
||
known state. When a change is detected, triggers _reload_mcp() and
|
||
informs the user so they know the tool list has been refreshed.
|
||
"""
|
||
import time
|
||
import yaml as _yaml
|
||
|
||
CONFIG_WATCH_INTERVAL = 5.0 # seconds between config.yaml stat() calls
|
||
|
||
now = time.monotonic()
|
||
if now - self._last_config_check < CONFIG_WATCH_INTERVAL:
|
||
return
|
||
self._last_config_check = now
|
||
|
||
from hermes_cli.config import get_config_path as _get_config_path
|
||
cfg_path = _get_config_path()
|
||
if not cfg_path.exists():
|
||
return
|
||
|
||
try:
|
||
mtime = cfg_path.stat().st_mtime
|
||
except OSError:
|
||
return
|
||
|
||
if mtime == self._config_mtime:
|
||
return # File unchanged — fast path
|
||
|
||
# File changed — check whether mcp_servers section changed
|
||
self._config_mtime = mtime
|
||
try:
|
||
with open(cfg_path, encoding="utf-8") as f:
|
||
new_cfg = _yaml.safe_load(f) or {}
|
||
except Exception:
|
||
return
|
||
|
||
new_mcp = new_cfg.get("mcp_servers") or {}
|
||
if new_mcp == self._config_mcp_servers:
|
||
return # mcp_servers unchanged (some other section was edited)
|
||
|
||
self._config_mcp_servers = new_mcp
|
||
# Notify user and reload
|
||
print()
|
||
print("🔄 MCP server config changed — reloading connections...")
|
||
with self._busy_command(self._slow_command_status("/reload-mcp")):
|
||
self._reload_mcp()
|
||
|
||
def _reload_mcp(self):
|
||
"""Reload MCP servers: disconnect all, re-read config.yaml, reconnect.
|
||
|
||
After reconnecting, refreshes the agent's tool list so the model
|
||
sees the updated tools on the next turn.
|
||
"""
|
||
try:
|
||
from tools.mcp_tool import shutdown_mcp_servers, discover_mcp_tools, _load_mcp_config, _servers, _lock
|
||
|
||
# Capture old server names
|
||
with _lock:
|
||
old_servers = set(_servers.keys())
|
||
|
||
if not self._command_running:
|
||
print("🔄 Reloading MCP servers...")
|
||
|
||
# Shutdown existing connections
|
||
shutdown_mcp_servers()
|
||
|
||
# Reconnect (reads config.yaml fresh)
|
||
new_tools = discover_mcp_tools()
|
||
|
||
# Compute what changed
|
||
with _lock:
|
||
connected_servers = set(_servers.keys())
|
||
|
||
added = connected_servers - old_servers
|
||
removed = old_servers - connected_servers
|
||
reconnected = connected_servers & old_servers
|
||
|
||
if reconnected:
|
||
print(f" ♻️ Reconnected: {', '.join(sorted(reconnected))}")
|
||
if added:
|
||
print(f" ➕ Added: {', '.join(sorted(added))}")
|
||
if removed:
|
||
print(f" ➖ Removed: {', '.join(sorted(removed))}")
|
||
if not connected_servers:
|
||
print(" No MCP servers connected.")
|
||
else:
|
||
print(f" 🔧 {len(new_tools)} tool(s) available from {len(connected_servers)} server(s)")
|
||
|
||
# Refresh the agent's tool list so the model can call new tools
|
||
if self.agent is not None:
|
||
from model_tools import get_tool_definitions
|
||
self.agent.tools = get_tool_definitions(
|
||
enabled_toolsets=self.agent.enabled_toolsets
|
||
if hasattr(self.agent, "enabled_toolsets") else None,
|
||
quiet_mode=True,
|
||
)
|
||
self.agent.valid_tool_names = {
|
||
tool["function"]["name"] for tool in self.agent.tools
|
||
} if self.agent.tools else set()
|
||
|
||
# Inject a message at the END of conversation history so the
|
||
# model knows tools changed. Appended after all existing
|
||
# messages to preserve prompt-cache for the prefix.
|
||
change_parts = []
|
||
if added:
|
||
change_parts.append(f"Added servers: {', '.join(sorted(added))}")
|
||
if removed:
|
||
change_parts.append(f"Removed servers: {', '.join(sorted(removed))}")
|
||
if reconnected:
|
||
change_parts.append(f"Reconnected servers: {', '.join(sorted(reconnected))}")
|
||
tool_summary = f"{len(new_tools)} MCP tool(s) now available" if new_tools else "No MCP tools available"
|
||
change_detail = ". ".join(change_parts) + ". " if change_parts else ""
|
||
self.conversation_history.append({
|
||
"role": "user",
|
||
"content": f"[SYSTEM: MCP servers have been reloaded. {change_detail}{tool_summary}. The tool list for this conversation has been updated accordingly.]",
|
||
})
|
||
|
||
# Persist session immediately so the session log reflects the
|
||
# updated tools list (self.agent.tools was refreshed above).
|
||
if self.agent is not None:
|
||
try:
|
||
self.agent._persist_session(
|
||
self.conversation_history,
|
||
self.conversation_history,
|
||
)
|
||
except Exception:
|
||
pass # Best-effort
|
||
|
||
print(f" ✅ Agent updated — {len(self.agent.tools if self.agent else [])} tool(s) available")
|
||
|
||
except Exception as e:
|
||
print(f" ❌ MCP reload failed: {e}")
|
||
|
||
# ====================================================================
|
||
# Tool progress callback (audio cues for voice mode)
|
||
# ====================================================================
|
||
|
||
def _on_tool_progress(self, function_name: str, preview: str, function_args: dict):
|
||
"""Called when a tool starts executing. Plays audio cue in voice mode."""
|
||
if not self._voice_mode:
|
||
return
|
||
# Skip internal/thinking tools
|
||
if function_name.startswith("_"):
|
||
return
|
||
try:
|
||
from tools.voice_mode import play_beep
|
||
# Short, subtle tick sound (higher pitch, very brief)
|
||
threading.Thread(
|
||
target=play_beep,
|
||
kwargs={"frequency": 1200, "duration": 0.06, "count": 1},
|
||
daemon=True,
|
||
).start()
|
||
except Exception:
|
||
pass
|
||
|
||
# ====================================================================
|
||
# Voice mode methods
|
||
# ====================================================================
|
||
|
||
def _voice_start_recording(self):
|
||
"""Start capturing audio from the microphone."""
|
||
if getattr(self, '_should_exit', False):
|
||
return
|
||
from tools.voice_mode import AudioRecorder, check_voice_requirements
|
||
|
||
reqs = check_voice_requirements()
|
||
if not reqs["audio_available"]:
|
||
raise RuntimeError(
|
||
"Voice mode requires sounddevice and numpy.\n"
|
||
"Install with: pip install sounddevice numpy\n"
|
||
"Or: pip install hermes-agent[voice]"
|
||
)
|
||
if not reqs.get("stt_available", reqs.get("stt_key_set")):
|
||
raise RuntimeError(
|
||
"Voice mode requires an STT provider for transcription.\n"
|
||
"Option 1: pip install faster-whisper (free, local)\n"
|
||
"Option 2: Set GROQ_API_KEY (free tier)\n"
|
||
"Option 3: Set VOICE_TOOLS_OPENAI_KEY (paid)"
|
||
)
|
||
|
||
# Prevent double-start from concurrent threads (atomic check-and-set)
|
||
with self._voice_lock:
|
||
if self._voice_recording:
|
||
return
|
||
self._voice_recording = True
|
||
|
||
# Load silence detection params from config
|
||
voice_cfg = {}
|
||
try:
|
||
from hermes_cli.config import load_config
|
||
voice_cfg = load_config().get("voice", {})
|
||
except Exception:
|
||
pass
|
||
|
||
if self._voice_recorder is None:
|
||
self._voice_recorder = AudioRecorder()
|
||
|
||
# Apply config-driven silence params
|
||
self._voice_recorder._silence_threshold = voice_cfg.get("silence_threshold", 200)
|
||
self._voice_recorder._silence_duration = voice_cfg.get("silence_duration", 3.0)
|
||
|
||
def _on_silence():
|
||
"""Called by AudioRecorder when silence is detected after speech."""
|
||
with self._voice_lock:
|
||
if not self._voice_recording:
|
||
return
|
||
_cprint(f"\n{_DIM}Silence detected, auto-stopping...{_RST}")
|
||
if hasattr(self, '_app') and self._app:
|
||
self._app.invalidate()
|
||
self._voice_stop_and_transcribe()
|
||
|
||
# Audio cue: single beep BEFORE starting stream (avoid CoreAudio conflict)
|
||
try:
|
||
from tools.voice_mode import play_beep
|
||
play_beep(frequency=880, count=1)
|
||
except Exception:
|
||
pass
|
||
|
||
try:
|
||
self._voice_recorder.start(on_silence_stop=_on_silence)
|
||
except Exception:
|
||
with self._voice_lock:
|
||
self._voice_recording = False
|
||
raise
|
||
_cprint(f"\n{_GOLD}● Recording...{_RST} {_DIM}(auto-stops on silence | Ctrl+B to stop & exit continuous){_RST}")
|
||
|
||
# Periodically refresh prompt to update audio level indicator
|
||
def _refresh_level():
|
||
while True:
|
||
with self._voice_lock:
|
||
still_recording = self._voice_recording
|
||
if not still_recording:
|
||
break
|
||
if hasattr(self, '_app') and self._app:
|
||
self._app.invalidate()
|
||
time.sleep(0.15)
|
||
threading.Thread(target=_refresh_level, daemon=True).start()
|
||
|
||
def _voice_stop_and_transcribe(self):
|
||
"""Stop recording, transcribe via STT, and queue the transcript as input."""
|
||
# Atomic guard: only one thread can enter stop-and-transcribe.
|
||
# Set _voice_processing immediately so concurrent Ctrl+B presses
|
||
# don't race into the START path while recorder.stop() holds its lock.
|
||
with self._voice_lock:
|
||
if not self._voice_recording:
|
||
return
|
||
self._voice_recording = False
|
||
self._voice_processing = True
|
||
|
||
submitted = False
|
||
wav_path = None
|
||
try:
|
||
if self._voice_recorder is None:
|
||
return
|
||
|
||
wav_path = self._voice_recorder.stop()
|
||
|
||
# Audio cue: double beep after stream stopped (no CoreAudio conflict)
|
||
try:
|
||
from tools.voice_mode import play_beep
|
||
play_beep(frequency=660, count=2)
|
||
except Exception:
|
||
pass
|
||
|
||
if wav_path is None:
|
||
_cprint(f"{_DIM}No speech detected.{_RST}")
|
||
return
|
||
|
||
# _voice_processing is already True (set atomically above)
|
||
if hasattr(self, '_app') and self._app:
|
||
self._app.invalidate()
|
||
_cprint(f"{_DIM}Transcribing...{_RST}")
|
||
|
||
# Get STT model from config
|
||
stt_model = None
|
||
try:
|
||
from hermes_cli.config import load_config
|
||
stt_config = load_config().get("stt", {})
|
||
stt_model = stt_config.get("model")
|
||
except Exception:
|
||
pass
|
||
|
||
from tools.voice_mode import transcribe_recording
|
||
result = transcribe_recording(wav_path, model=stt_model)
|
||
|
||
if result.get("success") and result.get("transcript", "").strip():
|
||
transcript = result["transcript"].strip()
|
||
self._pending_input.put(transcript)
|
||
submitted = True
|
||
elif result.get("success"):
|
||
_cprint(f"{_DIM}No speech detected.{_RST}")
|
||
else:
|
||
error = result.get("error", "Unknown error")
|
||
_cprint(f"\n{_DIM}Transcription failed: {error}{_RST}")
|
||
|
||
except Exception as e:
|
||
_cprint(f"\n{_DIM}Voice processing error: {e}{_RST}")
|
||
finally:
|
||
with self._voice_lock:
|
||
self._voice_processing = False
|
||
if hasattr(self, '_app') and self._app:
|
||
self._app.invalidate()
|
||
# Clean up temp file
|
||
try:
|
||
if wav_path and os.path.isfile(wav_path):
|
||
os.unlink(wav_path)
|
||
except Exception:
|
||
pass
|
||
|
||
# Track consecutive no-speech cycles to avoid infinite restart loops.
|
||
if not submitted:
|
||
self._no_speech_count = getattr(self, '_no_speech_count', 0) + 1
|
||
if self._no_speech_count >= 3:
|
||
self._voice_continuous = False
|
||
self._no_speech_count = 0
|
||
_cprint(f"{_DIM}No speech detected 3 times, continuous mode stopped.{_RST}")
|
||
return
|
||
else:
|
||
self._no_speech_count = 0
|
||
|
||
# If no transcript was submitted but continuous mode is active,
|
||
# restart recording so the user can keep talking.
|
||
# (When transcript IS submitted, process_loop handles restart
|
||
# after chat() completes.)
|
||
if self._voice_continuous and not submitted and not self._voice_recording:
|
||
def _restart_recording():
|
||
try:
|
||
self._voice_start_recording()
|
||
if hasattr(self, '_app') and self._app:
|
||
self._app.invalidate()
|
||
except Exception as e:
|
||
_cprint(f"{_DIM}Voice auto-restart failed: {e}{_RST}")
|
||
threading.Thread(target=_restart_recording, daemon=True).start()
|
||
|
||
def _voice_speak_response(self, text: str):
|
||
"""Speak the agent's response aloud using TTS (runs in background thread)."""
|
||
if not self._voice_tts:
|
||
return
|
||
self._voice_tts_done.clear()
|
||
try:
|
||
from tools.tts_tool import text_to_speech_tool
|
||
from tools.voice_mode import play_audio_file
|
||
import json
|
||
import re
|
||
|
||
# Strip markdown and non-speech content for cleaner TTS
|
||
tts_text = text[:4000] if len(text) > 4000 else text
|
||
tts_text = re.sub(r'```[\s\S]*?```', ' ', tts_text) # fenced code blocks
|
||
tts_text = re.sub(r'\[([^\]]+)\]\([^)]+\)', r'\1', tts_text) # [text](url) -> text
|
||
tts_text = re.sub(r'https?://\S+', '', tts_text) # URLs
|
||
tts_text = re.sub(r'\*\*(.+?)\*\*', r'\1', tts_text) # bold
|
||
tts_text = re.sub(r'\*(.+?)\*', r'\1', tts_text) # italic
|
||
tts_text = re.sub(r'`(.+?)`', r'\1', tts_text) # inline code
|
||
tts_text = re.sub(r'^#+\s*', '', tts_text, flags=re.MULTILINE) # headers
|
||
tts_text = re.sub(r'^\s*[-*]\s+', '', tts_text, flags=re.MULTILINE) # list items
|
||
tts_text = re.sub(r'---+', '', tts_text) # horizontal rules
|
||
tts_text = re.sub(r'\n{3,}', '\n\n', tts_text) # excessive newlines
|
||
tts_text = tts_text.strip()
|
||
if not tts_text:
|
||
return
|
||
|
||
# Use MP3 output for CLI playback (afplay doesn't handle OGG well).
|
||
# The TTS tool may auto-convert MP3->OGG, but the original MP3 remains.
|
||
os.makedirs(os.path.join(tempfile.gettempdir(), "hermes_voice"), exist_ok=True)
|
||
mp3_path = os.path.join(
|
||
tempfile.gettempdir(), "hermes_voice",
|
||
f"tts_{time.strftime('%Y%m%d_%H%M%S')}.mp3",
|
||
)
|
||
|
||
text_to_speech_tool(text=tts_text, output_path=mp3_path)
|
||
|
||
# Play the MP3 directly (the TTS tool returns OGG path but MP3 still exists)
|
||
if os.path.isfile(mp3_path) and os.path.getsize(mp3_path) > 0:
|
||
play_audio_file(mp3_path)
|
||
# Clean up
|
||
try:
|
||
os.unlink(mp3_path)
|
||
ogg_path = mp3_path.rsplit(".", 1)[0] + ".ogg"
|
||
if os.path.isfile(ogg_path):
|
||
os.unlink(ogg_path)
|
||
except OSError:
|
||
pass
|
||
except Exception as e:
|
||
logger.warning("Voice TTS playback failed: %s", e)
|
||
_cprint(f"{_DIM}TTS playback failed: {e}{_RST}")
|
||
finally:
|
||
self._voice_tts_done.set()
|
||
|
||
def _handle_voice_command(self, command: str):
|
||
"""Handle /voice [on|off|tts|status] command."""
|
||
parts = command.strip().split(maxsplit=1)
|
||
subcommand = parts[1].lower().strip() if len(parts) > 1 else ""
|
||
|
||
if subcommand == "on":
|
||
self._enable_voice_mode()
|
||
elif subcommand == "off":
|
||
self._disable_voice_mode()
|
||
elif subcommand == "tts":
|
||
self._toggle_voice_tts()
|
||
elif subcommand == "status":
|
||
self._show_voice_status()
|
||
elif subcommand == "":
|
||
# Toggle
|
||
if self._voice_mode:
|
||
self._disable_voice_mode()
|
||
else:
|
||
self._enable_voice_mode()
|
||
else:
|
||
_cprint(f"Unknown voice subcommand: {subcommand}")
|
||
_cprint("Usage: /voice [on|off|tts|status]")
|
||
|
||
def _enable_voice_mode(self):
|
||
"""Enable voice mode after checking requirements."""
|
||
if self._voice_mode:
|
||
_cprint(f"{_DIM}Voice mode is already enabled.{_RST}")
|
||
return
|
||
|
||
from tools.voice_mode import check_voice_requirements, detect_audio_environment
|
||
|
||
# Environment detection -- warn and block in incompatible environments
|
||
env_check = detect_audio_environment()
|
||
if not env_check["available"]:
|
||
_cprint(f"\n{_GOLD}Voice mode unavailable in this environment:{_RST}")
|
||
for warning in env_check["warnings"]:
|
||
_cprint(f" {_DIM}{warning}{_RST}")
|
||
return
|
||
|
||
reqs = check_voice_requirements()
|
||
if not reqs["available"]:
|
||
_cprint(f"\n{_GOLD}Voice mode requirements not met:{_RST}")
|
||
for line in reqs["details"].split("\n"):
|
||
_cprint(f" {_DIM}{line}{_RST}")
|
||
if reqs["missing_packages"]:
|
||
_cprint(f"\n {_BOLD}Install: pip install {' '.join(reqs['missing_packages'])}{_RST}")
|
||
_cprint(f" {_DIM}Or: pip install hermes-agent[voice]{_RST}")
|
||
return
|
||
|
||
with self._voice_lock:
|
||
self._voice_mode = True
|
||
|
||
# Check config for auto_tts
|
||
try:
|
||
from hermes_cli.config import load_config
|
||
voice_config = load_config().get("voice", {})
|
||
if voice_config.get("auto_tts", False):
|
||
with self._voice_lock:
|
||
self._voice_tts = True
|
||
except Exception:
|
||
pass
|
||
|
||
# Voice mode instruction is injected as a user message prefix (not a
|
||
# system prompt change) to avoid invalidating the prompt cache. See
|
||
# _voice_message_prefix property and its usage in _process_message().
|
||
|
||
tts_status = " (TTS enabled)" if self._voice_tts else ""
|
||
try:
|
||
from hermes_cli.config import load_config
|
||
_raw_ptt = load_config().get("voice", {}).get("record_key", "ctrl+b")
|
||
_ptt_key = _raw_ptt.lower().replace("ctrl+", "c-").replace("alt+", "a-")
|
||
except Exception:
|
||
_ptt_key = "c-b"
|
||
_ptt_display = _ptt_key.replace("c-", "Ctrl+").upper()
|
||
_cprint(f"\n{_GOLD}Voice mode enabled{tts_status}{_RST}")
|
||
_cprint(f" {_DIM}{_ptt_display} to start/stop recording{_RST}")
|
||
_cprint(f" {_DIM}/voice tts to toggle speech output{_RST}")
|
||
_cprint(f" {_DIM}/voice off to disable voice mode{_RST}")
|
||
|
||
def _disable_voice_mode(self):
|
||
"""Disable voice mode, cancel any active recording, and stop TTS."""
|
||
recorder = None
|
||
with self._voice_lock:
|
||
if self._voice_recording and self._voice_recorder:
|
||
self._voice_recorder.cancel()
|
||
self._voice_recording = False
|
||
recorder = self._voice_recorder
|
||
self._voice_mode = False
|
||
self._voice_tts = False
|
||
self._voice_continuous = False
|
||
|
||
# Shut down the persistent audio stream in background
|
||
if recorder is not None:
|
||
def _bg_shutdown(rec=recorder):
|
||
try:
|
||
rec.shutdown()
|
||
except Exception:
|
||
pass
|
||
threading.Thread(target=_bg_shutdown, daemon=True).start()
|
||
self._voice_recorder = None
|
||
|
||
# Stop any active TTS playback
|
||
try:
|
||
from tools.voice_mode import stop_playback
|
||
stop_playback()
|
||
except Exception:
|
||
pass
|
||
self._voice_tts_done.set()
|
||
|
||
_cprint(f"\n{_DIM}Voice mode disabled.{_RST}")
|
||
|
||
def _toggle_voice_tts(self):
|
||
"""Toggle TTS output for voice mode."""
|
||
if not self._voice_mode:
|
||
_cprint(f"{_DIM}Enable voice mode first: /voice on{_RST}")
|
||
return
|
||
|
||
with self._voice_lock:
|
||
self._voice_tts = not self._voice_tts
|
||
status = "enabled" if self._voice_tts else "disabled"
|
||
|
||
if self._voice_tts:
|
||
from tools.tts_tool import check_tts_requirements
|
||
if not check_tts_requirements():
|
||
_cprint(f"{_DIM}Warning: No TTS provider available. Install edge-tts or set API keys.{_RST}")
|
||
|
||
_cprint(f"{_GOLD}Voice TTS {status}.{_RST}")
|
||
|
||
def _show_voice_status(self):
|
||
"""Show current voice mode status."""
|
||
from hermes_cli.config import load_config
|
||
from tools.voice_mode import check_voice_requirements
|
||
|
||
reqs = check_voice_requirements()
|
||
|
||
_cprint(f"\n{_BOLD}Voice Mode Status{_RST}")
|
||
_cprint(f" Mode: {'ON' if self._voice_mode else 'OFF'}")
|
||
_cprint(f" TTS: {'ON' if self._voice_tts else 'OFF'}")
|
||
_cprint(f" Recording: {'YES' if self._voice_recording else 'no'}")
|
||
_raw_key = load_config().get("voice", {}).get("record_key", "ctrl+b")
|
||
_display_key = _raw_key.replace("ctrl+", "Ctrl+").upper() if "ctrl+" in _raw_key.lower() else _raw_key
|
||
_cprint(f" Record key: {_display_key}")
|
||
_cprint(f"\n {_BOLD}Requirements:{_RST}")
|
||
for line in reqs["details"].split("\n"):
|
||
_cprint(f" {line}")
|
||
|
||
def _clarify_callback(self, question, choices):
|
||
"""
|
||
Platform callback for the clarify tool. Called from the agent thread.
|
||
|
||
Sets up the interactive selection UI (or freetext prompt for open-ended
|
||
questions), then blocks until the user responds via the prompt_toolkit
|
||
key bindings. If no response arrives within the configured timeout the
|
||
question is dismissed and the agent is told to decide on its own.
|
||
"""
|
||
import time as _time
|
||
|
||
timeout = CLI_CONFIG.get("clarify", {}).get("timeout", 120)
|
||
response_queue = queue.Queue()
|
||
is_open_ended = not choices or len(choices) == 0
|
||
|
||
self._clarify_state = {
|
||
"question": question,
|
||
"choices": choices if not is_open_ended else [],
|
||
"selected": 0,
|
||
"response_queue": response_queue,
|
||
}
|
||
self._clarify_deadline = _time.monotonic() + timeout
|
||
# Open-ended questions skip straight to freetext input
|
||
self._clarify_freetext = is_open_ended
|
||
|
||
# Trigger prompt_toolkit repaint from this (non-main) thread
|
||
self._invalidate()
|
||
|
||
# Poll for the user's response. The countdown in the hint line
|
||
# updates on each invalidate — but frequent repaints cause visible
|
||
# flicker in some terminals (Kitty, ghostty). We only refresh the
|
||
# countdown every 5 s; selection changes (↑/↓) trigger instant
|
||
# Poll for the user's response. The countdown in the hint line
|
||
# updates on each invalidate — but frequent repaints cause visible
|
||
# flicker in some terminals (Kitty, ghostty). We only refresh the
|
||
# countdown every 5 s; selection changes (↑/↓) trigger instant
|
||
# repaints via the key bindings.
|
||
_last_countdown_refresh = _time.monotonic()
|
||
while True:
|
||
try:
|
||
result = response_queue.get(timeout=1)
|
||
self._clarify_deadline = 0
|
||
return result
|
||
except queue.Empty:
|
||
remaining = self._clarify_deadline - _time.monotonic()
|
||
if remaining <= 0:
|
||
break
|
||
# Only repaint every 5 s for the countdown — avoids flicker
|
||
now = _time.monotonic()
|
||
if now - _last_countdown_refresh >= 5.0:
|
||
_last_countdown_refresh = now
|
||
self._invalidate()
|
||
if now - _last_countdown_refresh >= 5.0:
|
||
_last_countdown_refresh = now
|
||
self._invalidate()
|
||
|
||
# Timed out — tear down the UI and let the agent decide
|
||
self._clarify_state = None
|
||
self._clarify_freetext = False
|
||
self._clarify_deadline = 0
|
||
self._invalidate()
|
||
_cprint(f"\n{_DIM}(clarify timed out after {timeout}s — agent will decide){_RST}")
|
||
return (
|
||
"The user did not provide a response within the time limit. "
|
||
"Use your best judgement to make the choice and proceed."
|
||
)
|
||
|
||
def _sudo_password_callback(self) -> str:
|
||
"""
|
||
Prompt for sudo password through the prompt_toolkit UI.
|
||
|
||
Called from the agent thread when a sudo command is encountered.
|
||
Uses the same clarify-style mechanism: sets UI state, waits on a
|
||
queue for the user's response via the Enter key binding.
|
||
"""
|
||
import time as _time
|
||
|
||
timeout = 45
|
||
response_queue = queue.Queue()
|
||
|
||
self._sudo_state = {
|
||
"response_queue": response_queue,
|
||
}
|
||
self._sudo_deadline = _time.monotonic() + timeout
|
||
|
||
self._invalidate()
|
||
|
||
while True:
|
||
try:
|
||
result = response_queue.get(timeout=1)
|
||
self._sudo_state = None
|
||
self._sudo_deadline = 0
|
||
self._invalidate()
|
||
if result:
|
||
_cprint(f"\n{_DIM} ✓ Password received (cached for session){_RST}")
|
||
else:
|
||
_cprint(f"\n{_DIM} ⏭ Skipped{_RST}")
|
||
return result
|
||
except queue.Empty:
|
||
remaining = self._sudo_deadline - _time.monotonic()
|
||
if remaining <= 0:
|
||
break
|
||
self._invalidate()
|
||
|
||
self._sudo_state = None
|
||
self._sudo_deadline = 0
|
||
self._invalidate()
|
||
_cprint(f"\n{_DIM} ⏱ Timeout — continuing without sudo{_RST}")
|
||
return ""
|
||
|
||
def _approval_callback(self, command: str, description: str,
|
||
*, allow_permanent: bool = True) -> str:
|
||
"""
|
||
Prompt for dangerous command approval through the prompt_toolkit UI.
|
||
|
||
Called from the agent thread. Shows a selection UI similar to clarify
|
||
with choices: once / session / always / deny. When allow_permanent
|
||
is False (tirith warnings present), the 'always' option is hidden.
|
||
Long commands also get a 'view' option so the full command can be
|
||
expanded before deciding.
|
||
|
||
Uses _approval_lock to serialize concurrent requests (e.g. from
|
||
parallel delegation subtasks) so each prompt gets its own turn
|
||
and the shared _approval_state / _approval_deadline aren't clobbered.
|
||
"""
|
||
import time as _time
|
||
|
||
with self._approval_lock:
|
||
timeout = 60
|
||
response_queue = queue.Queue()
|
||
|
||
self._approval_state = {
|
||
"command": command,
|
||
"description": description,
|
||
"choices": self._approval_choices(command, allow_permanent=allow_permanent),
|
||
"selected": 0,
|
||
"response_queue": response_queue,
|
||
}
|
||
self._approval_deadline = _time.monotonic() + timeout
|
||
|
||
self._invalidate()
|
||
|
||
_last_countdown_refresh = _time.monotonic()
|
||
while True:
|
||
try:
|
||
result = response_queue.get(timeout=1)
|
||
self._approval_state = None
|
||
self._approval_deadline = 0
|
||
self._invalidate()
|
||
return result
|
||
except queue.Empty:
|
||
remaining = self._approval_deadline - _time.monotonic()
|
||
if remaining <= 0:
|
||
break
|
||
now = _time.monotonic()
|
||
if now - _last_countdown_refresh >= 5.0:
|
||
_last_countdown_refresh = now
|
||
self._invalidate()
|
||
|
||
self._approval_state = None
|
||
self._approval_deadline = 0
|
||
self._invalidate()
|
||
_cprint(f"\n{_DIM} ⏱ Timeout — denying command{_RST}")
|
||
return "deny"
|
||
|
||
def _approval_choices(self, command: str, *, allow_permanent: bool = True) -> list[str]:
|
||
"""Return approval choices for a dangerous command prompt."""
|
||
choices = ["once", "session", "always", "deny"] if allow_permanent else ["once", "session", "deny"]
|
||
if len(command) > 70:
|
||
choices.append("view")
|
||
return choices
|
||
|
||
def _handle_approval_selection(self) -> None:
|
||
"""Process the currently selected dangerous-command approval choice."""
|
||
state = self._approval_state
|
||
if not state:
|
||
return
|
||
|
||
selected = state.get("selected", 0)
|
||
choices = state.get("choices") or []
|
||
if not (0 <= selected < len(choices)):
|
||
return
|
||
|
||
chosen = choices[selected]
|
||
if chosen == "view":
|
||
state["show_full"] = True
|
||
state["choices"] = [choice for choice in choices if choice != "view"]
|
||
if state["selected"] >= len(state["choices"]):
|
||
state["selected"] = max(0, len(state["choices"]) - 1)
|
||
self._invalidate()
|
||
return
|
||
|
||
state["response_queue"].put(chosen)
|
||
self._approval_state = None
|
||
self._invalidate()
|
||
|
||
def _get_approval_display_fragments(self):
|
||
"""Render the dangerous-command approval panel for the prompt_toolkit UI."""
|
||
state = self._approval_state
|
||
if not state:
|
||
return []
|
||
|
||
def _panel_box_width(title_text: str, content_lines: list[str], min_width: int = 46, max_width: int = 76) -> int:
|
||
term_cols = shutil.get_terminal_size((100, 20)).columns
|
||
longest = max([len(title_text)] + [len(line) for line in content_lines] + [min_width - 4])
|
||
inner = min(max(longest + 4, min_width - 2), max_width - 2, max(24, term_cols - 6))
|
||
return inner + 2
|
||
|
||
def _wrap_panel_text(text: str, width: int, subsequent_indent: str = "") -> list[str]:
|
||
wrapped = textwrap.wrap(
|
||
text,
|
||
width=max(8, width),
|
||
replace_whitespace=False,
|
||
drop_whitespace=False,
|
||
subsequent_indent=subsequent_indent,
|
||
)
|
||
return wrapped or [""]
|
||
|
||
def _append_panel_line(lines, border_style: str, content_style: str, text: str, box_width: int) -> None:
|
||
inner_width = max(0, box_width - 2)
|
||
lines.append((border_style, "│ "))
|
||
lines.append((content_style, text.ljust(inner_width)))
|
||
lines.append((border_style, " │\n"))
|
||
|
||
def _append_blank_panel_line(lines, border_style: str, box_width: int) -> None:
|
||
lines.append((border_style, "│" + (" " * box_width) + "│\n"))
|
||
|
||
command = state["command"]
|
||
description = state["description"]
|
||
choices = state["choices"]
|
||
selected = state.get("selected", 0)
|
||
show_full = state.get("show_full", False)
|
||
|
||
title = "⚠️ Dangerous Command"
|
||
cmd_display = command if show_full or len(command) <= 70 else command[:70] + '...'
|
||
choice_labels = {
|
||
"once": "Allow once",
|
||
"session": "Allow for this session",
|
||
"always": "Add to permanent allowlist",
|
||
"deny": "Deny",
|
||
"view": "Show full command",
|
||
}
|
||
|
||
preview_lines = _wrap_panel_text(description, 60)
|
||
preview_lines.extend(_wrap_panel_text(cmd_display, 60))
|
||
for i, choice in enumerate(choices):
|
||
prefix = '❯ ' if i == selected else ' '
|
||
preview_lines.extend(_wrap_panel_text(
|
||
f"{prefix}{choice_labels.get(choice, choice)}",
|
||
60,
|
||
subsequent_indent=" ",
|
||
))
|
||
|
||
box_width = _panel_box_width(title, preview_lines)
|
||
inner_text_width = max(8, box_width - 2)
|
||
|
||
lines = []
|
||
lines.append(('class:approval-border', '╭' + ('─' * box_width) + '╮\n'))
|
||
_append_panel_line(lines, 'class:approval-border', 'class:approval-title', title, box_width)
|
||
_append_blank_panel_line(lines, 'class:approval-border', box_width)
|
||
for wrapped in _wrap_panel_text(description, inner_text_width):
|
||
_append_panel_line(lines, 'class:approval-border', 'class:approval-desc', wrapped, box_width)
|
||
for wrapped in _wrap_panel_text(cmd_display, inner_text_width):
|
||
_append_panel_line(lines, 'class:approval-border', 'class:approval-cmd', wrapped, box_width)
|
||
_append_blank_panel_line(lines, 'class:approval-border', box_width)
|
||
for i, choice in enumerate(choices):
|
||
label = choice_labels.get(choice, choice)
|
||
style = 'class:approval-selected' if i == selected else 'class:approval-choice'
|
||
prefix = '❯ ' if i == selected else ' '
|
||
for wrapped in _wrap_panel_text(f"{prefix}{label}", inner_text_width, subsequent_indent=" "):
|
||
_append_panel_line(lines, 'class:approval-border', style, wrapped, box_width)
|
||
_append_blank_panel_line(lines, 'class:approval-border', box_width)
|
||
lines.append(('class:approval-border', '╰' + ('─' * box_width) + '╯\n'))
|
||
return lines
|
||
|
||
def _secret_capture_callback(self, var_name: str, prompt: str, metadata=None) -> dict:
|
||
return prompt_for_secret(self, var_name, prompt, metadata)
|
||
|
||
def _submit_secret_response(self, value: str) -> None:
|
||
if not self._secret_state:
|
||
return
|
||
self._secret_state["response_queue"].put(value)
|
||
self._secret_state = None
|
||
self._secret_deadline = 0
|
||
self._invalidate()
|
||
|
||
def _cancel_secret_capture(self) -> None:
|
||
self._submit_secret_response("")
|
||
|
||
def _clear_secret_input_buffer(self) -> None:
|
||
if getattr(self, "_app", None):
|
||
try:
|
||
self._app.current_buffer.reset()
|
||
except Exception:
|
||
pass
|
||
|
||
def _clear_current_input(self) -> None:
|
||
if getattr(self, "_app", None):
|
||
try:
|
||
self._app.current_buffer.text = ""
|
||
except Exception:
|
||
pass
|
||
|
||
|
||
def chat(self, message, images: list = None) -> Optional[str]:
|
||
"""
|
||
Send a message to the agent and get a response.
|
||
|
||
Handles streaming output, interrupt detection (user typing while agent
|
||
is working), and re-queueing of interrupted messages.
|
||
|
||
Uses a dedicated _interrupt_queue (separate from _pending_input) to avoid
|
||
race conditions between the process_loop and interrupt monitoring. Messages
|
||
typed while the agent is running go to _interrupt_queue; messages typed while
|
||
idle go to _pending_input.
|
||
|
||
Args:
|
||
message: The user's message (str or multimodal content list)
|
||
images: Optional list of Path objects for attached images
|
||
|
||
Returns:
|
||
The agent's response, or None on error
|
||
"""
|
||
# Single-query and direct chat callers do not go through run(), so
|
||
# register secure secret capture here as well.
|
||
set_secret_capture_callback(self._secret_capture_callback)
|
||
|
||
# Refresh provider credentials if needed (handles key rotation transparently)
|
||
if not self._ensure_runtime_credentials():
|
||
return None
|
||
|
||
turn_route = self._resolve_turn_agent_config(message)
|
||
if turn_route["signature"] != self._active_agent_route_signature:
|
||
self.agent = None
|
||
|
||
# Initialize agent if needed
|
||
if not self._init_agent(
|
||
model_override=turn_route["model"],
|
||
runtime_override=turn_route["runtime"],
|
||
route_label=turn_route["label"],
|
||
):
|
||
return None
|
||
|
||
# Pre-process images through the vision tool (Gemini Flash) so the
|
||
# main model receives text descriptions instead of raw base64 image
|
||
# content — works with any model, not just vision-capable ones.
|
||
if images:
|
||
message = self._preprocess_images_with_vision(
|
||
message if isinstance(message, str) else "", images
|
||
)
|
||
|
||
# Expand @ context references (e.g. @file:main.py, @diff, @folder:src/)
|
||
if isinstance(message, str) and "@" in message:
|
||
try:
|
||
from agent.context_references import preprocess_context_references
|
||
from agent.model_metadata import get_model_context_length
|
||
_ctx_len = get_model_context_length(
|
||
self.model, base_url=self.base_url or "", api_key=self.api_key or "")
|
||
_ctx_result = preprocess_context_references(
|
||
message, cwd=os.getcwd(), context_length=_ctx_len)
|
||
if _ctx_result.expanded or _ctx_result.blocked:
|
||
if _ctx_result.references:
|
||
_cprint(
|
||
f" {_DIM}[@ context: {len(_ctx_result.references)} ref(s), "
|
||
f"{_ctx_result.injected_tokens} tokens]{_RST}")
|
||
for w in _ctx_result.warnings:
|
||
_cprint(f" {_DIM}⚠ {w}{_RST}")
|
||
if _ctx_result.blocked:
|
||
return "\n".join(_ctx_result.warnings) or "Context injection refused."
|
||
message = _ctx_result.message
|
||
except Exception as e:
|
||
logging.debug("@ context reference expansion failed: %s", e)
|
||
|
||
# Add user message to history
|
||
self.conversation_history.append({"role": "user", "content": message})
|
||
|
||
ChatConsole().print(f"[{_accent_hex()}]{'─' * 40}[/]")
|
||
print(flush=True)
|
||
|
||
try:
|
||
# Run the conversation with interrupt monitoring
|
||
result = None
|
||
|
||
# Reset streaming display state for this turn
|
||
self._reset_stream_state()
|
||
|
||
# --- Streaming TTS setup ---
|
||
# When ElevenLabs is the TTS provider and sounddevice is available,
|
||
# we stream audio sentence-by-sentence as the agent generates tokens
|
||
# instead of waiting for the full response.
|
||
use_streaming_tts = False
|
||
_streaming_box_opened = False
|
||
text_queue = None
|
||
tts_thread = None
|
||
stream_callback = None
|
||
stop_event = None
|
||
|
||
if self._voice_tts:
|
||
try:
|
||
from tools.tts_tool import (
|
||
_load_tts_config as _load_tts_cfg,
|
||
_get_provider as _get_prov,
|
||
_import_elevenlabs,
|
||
_import_sounddevice,
|
||
stream_tts_to_speaker,
|
||
)
|
||
_tts_cfg = _load_tts_cfg()
|
||
if _get_prov(_tts_cfg) == "elevenlabs":
|
||
# Verify both ElevenLabs SDK and audio output are available
|
||
_import_elevenlabs()
|
||
_import_sounddevice()
|
||
use_streaming_tts = True
|
||
except (ImportError, OSError):
|
||
pass
|
||
except Exception:
|
||
pass
|
||
|
||
if use_streaming_tts:
|
||
text_queue = queue.Queue()
|
||
stop_event = threading.Event()
|
||
|
||
def display_callback(sentence: str):
|
||
"""Called by TTS consumer when a sentence is ready to display + speak."""
|
||
nonlocal _streaming_box_opened
|
||
if not _streaming_box_opened:
|
||
_streaming_box_opened = True
|
||
w = self.console.width
|
||
label = " ⚕ Hermes "
|
||
fill = w - 2 - len(label)
|
||
_cprint(f"\n{_GOLD}╭─{label}{'─' * max(fill - 1, 0)}╮{_RST}")
|
||
_cprint(sentence.rstrip())
|
||
|
||
tts_thread = threading.Thread(
|
||
target=stream_tts_to_speaker,
|
||
args=(text_queue, stop_event, self._voice_tts_done),
|
||
kwargs={"display_callback": display_callback},
|
||
daemon=True,
|
||
)
|
||
tts_thread.start()
|
||
|
||
def stream_callback(delta: str):
|
||
if text_queue is not None:
|
||
text_queue.put(delta)
|
||
|
||
# When voice mode is active, prepend a brief instruction so the
|
||
# model responds concisely. The prefix is API-call-local only —
|
||
# run_conversation persists the original clean user message.
|
||
_voice_prefix = ""
|
||
if self._voice_mode and isinstance(message, str):
|
||
_voice_prefix = (
|
||
"[Voice input — respond concisely and conversationally, "
|
||
"2-3 sentences max. No code blocks or markdown.] "
|
||
)
|
||
|
||
def run_agent():
|
||
nonlocal result
|
||
agent_message = _voice_prefix + message if _voice_prefix else message
|
||
result = self.agent.run_conversation(
|
||
user_message=agent_message,
|
||
conversation_history=self.conversation_history[:-1], # Exclude the message we just added
|
||
stream_callback=stream_callback,
|
||
task_id=self.session_id,
|
||
persist_user_message=message if _voice_prefix else None,
|
||
)
|
||
|
||
# Start agent in background thread
|
||
agent_thread = threading.Thread(target=run_agent)
|
||
agent_thread.start()
|
||
|
||
# Monitor the dedicated interrupt queue while the agent runs.
|
||
# _interrupt_queue is separate from _pending_input, so process_loop
|
||
# and chat() never compete for the same queue.
|
||
# When a clarify question is active, user input is handled entirely
|
||
# by the Enter key binding (routed to the clarify response queue),
|
||
# so we skip interrupt processing to avoid stealing that input.
|
||
interrupt_msg = None
|
||
while agent_thread.is_alive():
|
||
if hasattr(self, '_interrupt_queue'):
|
||
try:
|
||
interrupt_msg = self._interrupt_queue.get(timeout=0.1)
|
||
if interrupt_msg:
|
||
# If clarify is active, the Enter handler routes
|
||
# input directly; this queue shouldn't have anything.
|
||
# But if it does (race condition), don't interrupt.
|
||
if self._clarify_state or self._clarify_freetext:
|
||
continue
|
||
print(f"\n⚡ New message detected, interrupting...")
|
||
# Signal TTS to stop on interrupt
|
||
if stop_event is not None:
|
||
stop_event.set()
|
||
self.agent.interrupt(interrupt_msg)
|
||
# Debug: log to file (stdout may be devnull from redirect_stdout)
|
||
try:
|
||
_dbg = _hermes_home / "interrupt_debug.log"
|
||
with open(_dbg, "a") as _f:
|
||
import time as _t
|
||
_f.write(f"{_t.strftime('%H:%M:%S')} interrupt fired: msg={str(interrupt_msg)[:60]!r}, "
|
||
f"children={len(self.agent._active_children)}, "
|
||
f"parent._interrupt={self.agent._interrupt_requested}\n")
|
||
for _ci, _ch in enumerate(self.agent._active_children):
|
||
_f.write(f" child[{_ci}]._interrupt={_ch._interrupt_requested}\n")
|
||
except Exception:
|
||
pass
|
||
break
|
||
except queue.Empty:
|
||
# Force prompt_toolkit to flush any pending stdout
|
||
# output from the agent thread. Without this, the
|
||
# StdoutProxy buffer only flushes on renderer passes
|
||
# triggered by input events — on macOS this causes
|
||
# the CLI to appear frozen until the user types. (#1624)
|
||
self._invalidate(min_interval=0.15)
|
||
else:
|
||
# Fallback for non-interactive mode (e.g., single-query)
|
||
agent_thread.join(0.1)
|
||
|
||
agent_thread.join() # Ensure agent thread completes
|
||
|
||
# Flush any remaining streamed text and close the box
|
||
self._flush_stream()
|
||
|
||
# Signal end-of-text to TTS consumer and wait for it to finish
|
||
if use_streaming_tts and text_queue is not None:
|
||
text_queue.put(None) # sentinel
|
||
if tts_thread is not None:
|
||
tts_thread.join(timeout=120)
|
||
|
||
# Drain any remaining agent output still in the StdoutProxy
|
||
# buffer so tool/status lines render ABOVE our response box.
|
||
# The flush pushes data into the renderer queue; the short
|
||
# sleep lets the renderer actually paint it before we draw.
|
||
import time as _time
|
||
sys.stdout.flush()
|
||
_time.sleep(0.15)
|
||
|
||
# Update history with full conversation
|
||
self.conversation_history = result.get("messages", self.conversation_history) if result else self.conversation_history
|
||
|
||
# Get the final response
|
||
response = result.get("final_response", "") if result else ""
|
||
|
||
# Auto-generate session title after first exchange (non-blocking)
|
||
if response and result and not result.get("failed") and not result.get("partial"):
|
||
try:
|
||
from agent.title_generator import maybe_auto_title
|
||
maybe_auto_title(
|
||
self._session_db,
|
||
self.session_id,
|
||
message,
|
||
response,
|
||
self.conversation_history,
|
||
)
|
||
except Exception:
|
||
pass
|
||
|
||
# Handle failed or partial results (e.g., non-retryable errors, rate limits,
|
||
# truncated output, invalid tool calls). Both "failed" and "partial" with
|
||
# an empty final_response mean the agent couldn't produce a usable answer.
|
||
if result and (result.get("failed") or result.get("partial")) and not response:
|
||
error_detail = result.get("error", "Unknown error")
|
||
response = f"Error: {error_detail}"
|
||
# Stop continuous voice mode on persistent errors (e.g. 429 rate limit)
|
||
# to avoid an infinite error → record → error loop
|
||
if self._voice_continuous:
|
||
self._voice_continuous = False
|
||
_cprint(f"\n{_DIM}Continuous voice mode stopped due to error.{_RST}")
|
||
|
||
# Handle interrupt - check if we were interrupted
|
||
pending_message = None
|
||
if result and result.get("interrupted"):
|
||
pending_message = result.get("interrupt_message") or interrupt_msg
|
||
# Add indicator that we were interrupted
|
||
if response and pending_message:
|
||
response = response + "\n\n---\n_[Interrupted - processing new message]_"
|
||
|
||
response_previewed = result.get("response_previewed", False) if result else False
|
||
|
||
# Display reasoning (thinking) box if enabled and available.
|
||
# Skip when streaming already showed reasoning live.
|
||
if self.show_reasoning and result and not self._stream_started:
|
||
reasoning = result.get("last_reasoning")
|
||
if reasoning:
|
||
w = shutil.get_terminal_size().columns
|
||
r_label = " Reasoning "
|
||
r_fill = w - 2 - len(r_label)
|
||
r_top = f"{_DIM}┌─{r_label}{'─' * max(r_fill - 1, 0)}┐{_RST}"
|
||
r_bot = f"{_DIM}└{'─' * (w - 2)}┘{_RST}"
|
||
# Collapse long reasoning: show first 10 lines
|
||
lines = reasoning.strip().splitlines()
|
||
if len(lines) > 10:
|
||
display_reasoning = "\n".join(lines[:10])
|
||
display_reasoning += f"\n{_DIM} ... ({len(lines) - 10} more lines){_RST}"
|
||
else:
|
||
display_reasoning = reasoning.strip()
|
||
_cprint(f"\n{r_top}\n{_DIM}{display_reasoning}{_RST}\n{r_bot}")
|
||
|
||
if response and not response_previewed:
|
||
# Use skin engine for label/color with fallback
|
||
try:
|
||
from hermes_cli.skin_engine import get_active_skin
|
||
_skin = get_active_skin()
|
||
label = _skin.get_branding("response_label", "⚕ Hermes")
|
||
_resp_color = _skin.get_color("response_border", "#CD7F32")
|
||
_resp_text = _skin.get_color("banner_text", "#FFF8DC")
|
||
except Exception:
|
||
label = "⚕ Hermes"
|
||
_resp_color = "#CD7F32"
|
||
_resp_text = "#FFF8DC"
|
||
|
||
is_error_response = result and (result.get("failed") or result.get("partial"))
|
||
already_streamed = self._stream_started and self._stream_box_opened and not is_error_response
|
||
if use_streaming_tts and _streaming_box_opened and not is_error_response:
|
||
# Text was already printed sentence-by-sentence; just close the box
|
||
w = shutil.get_terminal_size().columns
|
||
_cprint(f"\n{_GOLD}╰{'─' * (w - 2)}╯{_RST}")
|
||
elif already_streamed:
|
||
# Response was already streamed token-by-token with box framing;
|
||
# _flush_stream() already closed the box. Skip Rich Panel.
|
||
pass
|
||
else:
|
||
_chat_console = ChatConsole()
|
||
_chat_console.print(Panel(
|
||
_rich_text_from_ansi(response),
|
||
title=f"[{_resp_color} bold]{label}[/]",
|
||
title_align="left",
|
||
border_style=_resp_color,
|
||
style=_resp_text,
|
||
box=rich_box.HORIZONTALS,
|
||
padding=(1, 2),
|
||
))
|
||
|
||
|
||
# Play terminal bell when agent finishes (if enabled).
|
||
# Works over SSH — the bell propagates to the user's terminal.
|
||
if self.bell_on_complete:
|
||
sys.stdout.write("\a")
|
||
sys.stdout.flush()
|
||
|
||
# Speak response aloud if voice TTS is enabled
|
||
# Skip batch TTS when streaming TTS already handled it
|
||
if self._voice_tts and response and not use_streaming_tts:
|
||
threading.Thread(
|
||
target=self._voice_speak_response,
|
||
args=(response,),
|
||
daemon=True,
|
||
).start()
|
||
|
||
|
||
# Combine all interrupt messages (user may have typed multiple while waiting)
|
||
# and re-queue as one prompt for process_loop
|
||
if pending_message and hasattr(self, '_pending_input'):
|
||
all_parts = [pending_message]
|
||
while not self._interrupt_queue.empty():
|
||
try:
|
||
extra = self._interrupt_queue.get_nowait()
|
||
if extra:
|
||
all_parts.append(extra)
|
||
except queue.Empty:
|
||
break
|
||
combined = "\n".join(all_parts)
|
||
print(f"\n📨 Queued: '{combined[:50]}{'...' if len(combined) > 50 else ''}'")
|
||
self._pending_input.put(combined)
|
||
|
||
return response
|
||
|
||
except Exception as e:
|
||
print(f"Error: {e}")
|
||
return None
|
||
finally:
|
||
# Ensure streaming TTS resources are cleaned up even on error.
|
||
# Normal path sends the sentinel at line ~3568; this is a safety
|
||
# net for exception paths that skip it. Duplicate sentinels are
|
||
# harmless — stream_tts_to_speaker exits on the first None.
|
||
if text_queue is not None:
|
||
try:
|
||
text_queue.put_nowait(None)
|
||
except Exception:
|
||
pass
|
||
if stop_event is not None:
|
||
stop_event.set()
|
||
if tts_thread is not None and tts_thread.is_alive():
|
||
tts_thread.join(timeout=5)
|
||
|
||
def _print_exit_summary(self):
|
||
"""Print session resume info on exit, similar to Claude Code."""
|
||
print()
|
||
msg_count = len(self.conversation_history)
|
||
if msg_count > 0:
|
||
user_msgs = len([m for m in self.conversation_history if m.get("role") == "user"])
|
||
tool_calls = len([m for m in self.conversation_history if m.get("role") == "tool" or m.get("tool_calls")])
|
||
elapsed = datetime.now() - self.session_start
|
||
hours, remainder = divmod(int(elapsed.total_seconds()), 3600)
|
||
minutes, seconds = divmod(remainder, 60)
|
||
if hours > 0:
|
||
duration_str = f"{hours}h {minutes}m {seconds}s"
|
||
elif minutes > 0:
|
||
duration_str = f"{minutes}m {seconds}s"
|
||
else:
|
||
duration_str = f"{seconds}s"
|
||
|
||
print(f"Resume this session with:")
|
||
print(f" hermes --resume {self.session_id}")
|
||
print()
|
||
print(f"Session: {self.session_id}")
|
||
print(f"Duration: {duration_str}")
|
||
print(f"Messages: {msg_count} ({user_msgs} user, {tool_calls} tool calls)")
|
||
else:
|
||
try:
|
||
from hermes_cli.skin_engine import get_active_goodbye
|
||
goodbye = get_active_goodbye("Goodbye! ⚕")
|
||
except Exception:
|
||
goodbye = "Goodbye! ⚕"
|
||
print(goodbye)
|
||
|
||
def _get_tui_prompt_symbols(self) -> tuple[str, str]:
|
||
"""Return ``(normal_prompt, state_suffix)`` for the active skin.
|
||
|
||
``normal_prompt`` is the full ``branding.prompt_symbol``.
|
||
``state_suffix`` is what special states (sudo/secret/approval/agent)
|
||
should render after their leading icon.
|
||
"""
|
||
try:
|
||
from hermes_cli.skin_engine import get_active_prompt_symbol
|
||
symbol = get_active_prompt_symbol("❯ ")
|
||
except Exception:
|
||
symbol = "❯ "
|
||
|
||
symbol = (symbol or "❯ ").rstrip() + " "
|
||
stripped = symbol.rstrip()
|
||
if not stripped:
|
||
return "❯ ", "❯ "
|
||
|
||
parts = stripped.split()
|
||
candidate = parts[-1] if parts else ""
|
||
arrow_chars = ("❯", ">", "$", "#", "›", "»", "→")
|
||
if any(ch in candidate for ch in arrow_chars):
|
||
return symbol, candidate.rstrip() + " "
|
||
|
||
# Icon-only custom prompts should still remain visible in special states.
|
||
return symbol, symbol
|
||
|
||
def _audio_level_bar(self) -> str:
|
||
"""Return a visual audio level indicator based on current RMS."""
|
||
_LEVEL_BARS = " ▁▂▃▄▅▆▇"
|
||
rec = getattr(self, "_voice_recorder", None)
|
||
if rec is None:
|
||
return ""
|
||
rms = rec.current_rms
|
||
# Normalize RMS (0-32767) to 0-7 index, with log-ish scaling
|
||
# Typical speech RMS is 500-5000, we cap display at ~8000
|
||
level = min(rms, 8000) * 7 // 8000
|
||
return _LEVEL_BARS[level]
|
||
|
||
def _get_tui_prompt_fragments(self):
|
||
"""Return the prompt_toolkit fragments for the current interactive state."""
|
||
symbol, state_suffix = self._get_tui_prompt_symbols()
|
||
if self._voice_recording:
|
||
bar = self._audio_level_bar()
|
||
return [("class:voice-recording", f"● {bar} {state_suffix}")]
|
||
if self._voice_processing:
|
||
return [("class:voice-processing", f"◉ {state_suffix}")]
|
||
if self._sudo_state:
|
||
return [("class:sudo-prompt", f"🔐 {state_suffix}")]
|
||
if self._secret_state:
|
||
return [("class:sudo-prompt", f"🔑 {state_suffix}")]
|
||
if self._approval_state:
|
||
return [("class:prompt-working", f"⚠ {state_suffix}")]
|
||
if self._clarify_freetext:
|
||
return [("class:clarify-selected", f"✎ {state_suffix}")]
|
||
if self._clarify_state:
|
||
return [("class:prompt-working", f"? {state_suffix}")]
|
||
if self._command_running:
|
||
return [("class:prompt-working", f"{self._command_spinner_frame()} {state_suffix}")]
|
||
if self._agent_running:
|
||
return [("class:prompt-working", f"⚕ {state_suffix}")]
|
||
if self._voice_mode:
|
||
return [("class:voice-prompt", f"🎤 {state_suffix}")]
|
||
return [("class:prompt", symbol)]
|
||
|
||
def _get_tui_prompt_text(self) -> str:
|
||
"""Return the visible prompt text for width calculations."""
|
||
return "".join(text for _, text in self._get_tui_prompt_fragments())
|
||
|
||
def _build_tui_style_dict(self) -> dict[str, str]:
|
||
"""Layer the active skin's prompt_toolkit colors over the base TUI style."""
|
||
style_dict = dict(getattr(self, "_tui_style_base", {}) or {})
|
||
try:
|
||
from hermes_cli.skin_engine import get_prompt_toolkit_style_overrides
|
||
style_dict.update(get_prompt_toolkit_style_overrides())
|
||
except Exception:
|
||
pass
|
||
return style_dict
|
||
|
||
def _apply_tui_skin_style(self) -> bool:
|
||
"""Refresh prompt_toolkit styling for a running interactive TUI."""
|
||
if not getattr(self, "_app", None) or not getattr(self, "_tui_style_base", None):
|
||
return False
|
||
self._app.style = PTStyle.from_dict(self._build_tui_style_dict())
|
||
self._invalidate(min_interval=0.0)
|
||
return True
|
||
|
||
def run(self):
|
||
"""Run the interactive CLI loop with persistent input at bottom."""
|
||
self.show_banner()
|
||
|
||
# One-line Honcho session indicator (TTY-only, not captured by agent).
|
||
# Only show when the user explicitly configured Honcho for Hermes
|
||
# (not auto-enabled from a stray HONCHO_API_KEY env var).
|
||
try:
|
||
from honcho_integration.client import HonchoClientConfig
|
||
from agent.display import honcho_session_line, write_tty
|
||
hcfg = HonchoClientConfig.from_global_config()
|
||
if hcfg.enabled and hcfg.api_key and hcfg.explicitly_configured:
|
||
sname = hcfg.resolve_session_name(session_id=self.session_id)
|
||
if sname:
|
||
write_tty(honcho_session_line(hcfg.workspace_id, sname) + "\n")
|
||
except Exception:
|
||
pass
|
||
|
||
# If resuming a session, load history and display it immediately
|
||
# so the user has context before typing their first message.
|
||
if self._resumed:
|
||
if self._preload_resumed_session():
|
||
self._display_resumed_history()
|
||
|
||
try:
|
||
from hermes_cli.skin_engine import get_active_skin
|
||
_welcome_skin = get_active_skin()
|
||
_welcome_text = _welcome_skin.get_branding("welcome", "Welcome to Hermes Agent! Type your message or /help for commands.")
|
||
_welcome_color = _welcome_skin.get_color("banner_text", "#FFF8DC")
|
||
except Exception:
|
||
_welcome_text = "Welcome to Hermes Agent! Type your message or /help for commands."
|
||
_welcome_color = "#FFF8DC"
|
||
self.console.print(f"[{_welcome_color}]{_welcome_text}[/]")
|
||
self.console.print()
|
||
|
||
# State for async operation
|
||
self._agent_running = False
|
||
self._pending_input = queue.Queue() # For normal input (commands + new queries)
|
||
self._interrupt_queue = queue.Queue() # For messages typed while agent is running
|
||
self._should_exit = False
|
||
self._last_ctrl_c_time = 0 # Track double Ctrl+C for force exit
|
||
# Config file watcher — detect mcp_servers changes and auto-reload
|
||
from hermes_cli.config import get_config_path as _get_config_path
|
||
_cfg_path = _get_config_path()
|
||
self._config_mtime: float = _cfg_path.stat().st_mtime if _cfg_path.exists() else 0.0
|
||
self._config_mcp_servers: dict = self.config.get("mcp_servers") or {}
|
||
self._last_config_check: float = 0.0 # monotonic time of last check
|
||
|
||
# Clarify tool state: interactive question/answer with the user.
|
||
# When the agent calls the clarify tool, _clarify_state is set and
|
||
# the prompt_toolkit UI switches to a selection mode.
|
||
self._clarify_state = None # dict with question, choices, selected, response_queue
|
||
self._clarify_freetext = False # True when user chose "Other" and is typing
|
||
self._clarify_deadline = 0 # monotonic timestamp when the clarify times out
|
||
|
||
# Sudo password prompt state (similar mechanism to clarify)
|
||
self._sudo_state = None # dict with response_queue when active
|
||
self._sudo_deadline = 0
|
||
|
||
# Dangerous command approval state (similar mechanism to clarify)
|
||
self._approval_state = None # dict with command, description, choices, selected, response_queue
|
||
self._approval_deadline = 0
|
||
self._approval_lock = threading.Lock() # serialize concurrent approval prompts (delegation race fix)
|
||
|
||
# Slash command loading state
|
||
self._command_running = False
|
||
self._command_status = ""
|
||
|
||
# Secure secret capture state for skill setup
|
||
self._secret_state = None # dict with var_name, prompt, metadata, response_queue
|
||
self._secret_deadline = 0
|
||
|
||
# Clipboard image attachments (paste images into the CLI)
|
||
self._attached_images: list[Path] = []
|
||
self._image_counter = 0
|
||
|
||
# Voice mode state (protected by _voice_lock for cross-thread access)
|
||
self._voice_lock = threading.Lock()
|
||
self._voice_mode = False # Whether voice mode is enabled
|
||
self._voice_tts = False # Whether TTS output is enabled
|
||
self._voice_recorder = None # AudioRecorder instance (lazy init)
|
||
self._voice_recording = False # Whether currently recording
|
||
self._voice_processing = False # Whether STT is in progress
|
||
self._voice_continuous = False # Whether to auto-restart after agent responds
|
||
self._voice_tts_done = threading.Event() # Signals TTS playback finished
|
||
self._voice_tts_done.set() # Initially "done" (no TTS pending)
|
||
|
||
# Register callbacks so terminal_tool prompts route through our UI
|
||
set_sudo_password_callback(self._sudo_password_callback)
|
||
set_approval_callback(self._approval_callback)
|
||
set_secret_capture_callback(self._secret_capture_callback)
|
||
|
||
# Ensure tirith security scanner is available (downloads if needed)
|
||
try:
|
||
from tools.tirith_security import ensure_installed
|
||
ensure_installed(log_failures=False)
|
||
except Exception:
|
||
pass # Non-fatal — fail-open at scan time if unavailable
|
||
|
||
# Key bindings for the input area
|
||
kb = KeyBindings()
|
||
|
||
@kb.add('enter')
|
||
def handle_enter(event):
|
||
"""Handle Enter key - submit input.
|
||
|
||
Routes to the correct queue based on active UI state:
|
||
- Sudo password prompt: password goes to sudo response queue
|
||
- Approval selection: selected choice goes to approval response queue
|
||
- Clarify freetext mode: answer goes to the clarify response queue
|
||
- Clarify choice mode: selected choice goes to the clarify response queue
|
||
- Agent running: goes to _interrupt_queue (chat() monitors this)
|
||
- Agent idle: goes to _pending_input (process_loop monitors this)
|
||
Commands (starting with /) always go to _pending_input so they're
|
||
handled as commands, not sent as interrupt text to the agent.
|
||
"""
|
||
# --- Sudo password prompt: submit the typed password ---
|
||
if self._sudo_state:
|
||
text = event.app.current_buffer.text
|
||
self._sudo_state["response_queue"].put(text)
|
||
self._sudo_state = None
|
||
event.app.current_buffer.reset()
|
||
event.app.invalidate()
|
||
return
|
||
|
||
# --- Secret prompt: submit the typed secret ---
|
||
if self._secret_state:
|
||
text = event.app.current_buffer.text
|
||
self._submit_secret_response(text)
|
||
event.app.current_buffer.reset()
|
||
event.app.invalidate()
|
||
return
|
||
|
||
# --- Approval selection: confirm the highlighted choice ---
|
||
if self._approval_state:
|
||
self._handle_approval_selection()
|
||
event.app.invalidate()
|
||
return
|
||
|
||
# --- Clarify freetext mode: user typed their own answer ---
|
||
if self._clarify_freetext and self._clarify_state:
|
||
text = event.app.current_buffer.text.strip()
|
||
if text:
|
||
self._clarify_state["response_queue"].put(text)
|
||
self._clarify_state = None
|
||
self._clarify_freetext = False
|
||
event.app.current_buffer.reset()
|
||
event.app.invalidate()
|
||
return
|
||
|
||
# --- Clarify choice mode: confirm the highlighted selection ---
|
||
if self._clarify_state and not self._clarify_freetext:
|
||
state = self._clarify_state
|
||
selected = state["selected"]
|
||
choices = state.get("choices") or []
|
||
if selected < len(choices):
|
||
state["response_queue"].put(choices[selected])
|
||
self._clarify_state = None
|
||
event.app.invalidate()
|
||
else:
|
||
# "Other" selected → switch to freetext
|
||
self._clarify_freetext = True
|
||
event.app.invalidate()
|
||
return
|
||
|
||
# --- Normal input routing ---
|
||
text = event.app.current_buffer.text.strip()
|
||
has_images = bool(self._attached_images)
|
||
if text or has_images:
|
||
# Snapshot and clear attached images
|
||
images = list(self._attached_images)
|
||
self._attached_images.clear()
|
||
event.app.invalidate()
|
||
# Bundle text + images as a tuple when images are present
|
||
payload = (text, images) if images else text
|
||
if self._agent_running and not (text and text.startswith("/")):
|
||
self._interrupt_queue.put(payload)
|
||
# Debug: log to file when message enters interrupt queue
|
||
try:
|
||
_dbg = _hermes_home / "interrupt_debug.log"
|
||
with open(_dbg, "a") as _f:
|
||
import time as _t
|
||
_f.write(f"{_t.strftime('%H:%M:%S')} ENTER: queued interrupt msg={str(payload)[:60]!r}, "
|
||
f"agent_running={self._agent_running}\n")
|
||
except Exception:
|
||
pass
|
||
else:
|
||
self._pending_input.put(payload)
|
||
event.app.current_buffer.reset(append_to_history=True)
|
||
|
||
@kb.add('escape', 'enter')
|
||
def handle_alt_enter(event):
|
||
"""Alt+Enter inserts a newline for multi-line input."""
|
||
event.current_buffer.insert_text('\n')
|
||
|
||
@kb.add('c-j')
|
||
def handle_ctrl_enter(event):
|
||
"""Ctrl+Enter (c-j) inserts a newline. Most terminals send c-j for Ctrl+Enter."""
|
||
event.current_buffer.insert_text('\n')
|
||
|
||
@kb.add('tab', eager=True)
|
||
def handle_tab(event):
|
||
"""Tab: accept completion, auto-suggestion, or start completions.
|
||
|
||
Priority:
|
||
1. Completion menu open → accept selected completion
|
||
2. Ghost text suggestion available → accept auto-suggestion
|
||
3. Otherwise → start completion menu
|
||
|
||
After accepting a provider like 'anthropic:', the completion menu
|
||
closes and complete_while_typing doesn't fire (no keystroke).
|
||
This binding re-triggers completions so stage-2 models appear
|
||
immediately.
|
||
"""
|
||
buf = event.current_buffer
|
||
if buf.complete_state:
|
||
# Completion menu is open — accept the selection
|
||
completion = buf.complete_state.current_completion
|
||
if completion is None:
|
||
# Menu open but nothing selected — select first then grab it
|
||
buf.go_to_completion(0)
|
||
completion = buf.complete_state and buf.complete_state.current_completion
|
||
if completion is None:
|
||
return
|
||
# Accept the selected completion
|
||
buf.apply_completion(completion)
|
||
# If text now looks like "/model provider:", re-trigger completions
|
||
text = buf.document.text_before_cursor
|
||
if text.startswith("/model ") and text.endswith(":"):
|
||
buf.start_completion()
|
||
elif buf.suggestion and buf.suggestion.text:
|
||
# No completion menu, but there's a ghost text auto-suggestion — accept it
|
||
buf.insert_text(buf.suggestion.text)
|
||
else:
|
||
# No menu and no suggestion — start completions from scratch
|
||
buf.start_completion()
|
||
|
||
# --- Clarify tool: arrow-key navigation for multiple-choice questions ---
|
||
|
||
@kb.add('up', filter=Condition(lambda: bool(self._clarify_state) and not self._clarify_freetext))
|
||
def clarify_up(event):
|
||
"""Move selection up in clarify choices."""
|
||
if self._clarify_state:
|
||
self._clarify_state["selected"] = max(0, self._clarify_state["selected"] - 1)
|
||
event.app.invalidate()
|
||
|
||
@kb.add('down', filter=Condition(lambda: bool(self._clarify_state) and not self._clarify_freetext))
|
||
def clarify_down(event):
|
||
"""Move selection down in clarify choices."""
|
||
if self._clarify_state:
|
||
choices = self._clarify_state.get("choices") or []
|
||
max_idx = len(choices) # last index is the "Other" option
|
||
self._clarify_state["selected"] = min(max_idx, self._clarify_state["selected"] + 1)
|
||
event.app.invalidate()
|
||
|
||
# --- Dangerous command approval: arrow-key navigation ---
|
||
|
||
@kb.add('up', filter=Condition(lambda: bool(self._approval_state)))
|
||
def approval_up(event):
|
||
if self._approval_state:
|
||
self._approval_state["selected"] = max(0, self._approval_state["selected"] - 1)
|
||
event.app.invalidate()
|
||
|
||
@kb.add('down', filter=Condition(lambda: bool(self._approval_state)))
|
||
def approval_down(event):
|
||
if self._approval_state:
|
||
max_idx = len(self._approval_state["choices"]) - 1
|
||
self._approval_state["selected"] = min(max_idx, self._approval_state["selected"] + 1)
|
||
event.app.invalidate()
|
||
|
||
# --- History navigation: up/down browse history in normal input mode ---
|
||
# The TextArea is multiline, so by default up/down only move the cursor.
|
||
# Buffer.auto_up/auto_down handle both: cursor movement when multi-line,
|
||
# history browsing when on the first/last line (or single-line input).
|
||
_normal_input = Condition(
|
||
lambda: not self._clarify_state and not self._approval_state and not self._sudo_state and not self._secret_state
|
||
)
|
||
|
||
@kb.add('up', filter=_normal_input)
|
||
def history_up(event):
|
||
"""Up arrow: browse history when on first line, else move cursor up."""
|
||
event.app.current_buffer.auto_up(count=event.arg)
|
||
|
||
@kb.add('down', filter=_normal_input)
|
||
def history_down(event):
|
||
"""Down arrow: browse history when on last line, else move cursor down."""
|
||
event.app.current_buffer.auto_down(count=event.arg)
|
||
|
||
@kb.add('c-c')
|
||
def handle_ctrl_c(event):
|
||
"""Handle Ctrl+C - cancel interactive prompts, interrupt agent, or exit.
|
||
|
||
Priority:
|
||
0. Cancel active voice recording
|
||
1. Cancel active sudo/approval/clarify prompt
|
||
2. Interrupt the running agent (first press)
|
||
3. Force exit (second press within 2s, or when idle)
|
||
"""
|
||
import time as _time
|
||
now = _time.time()
|
||
|
||
# Cancel active voice recording.
|
||
# Run cancel() in a background thread to prevent blocking the
|
||
# event loop if AudioRecorder._lock or CoreAudio takes time.
|
||
_should_cancel_voice = False
|
||
_recorder_ref = None
|
||
with cli_ref._voice_lock:
|
||
if cli_ref._voice_recording and cli_ref._voice_recorder:
|
||
_recorder_ref = cli_ref._voice_recorder
|
||
cli_ref._voice_recording = False
|
||
cli_ref._voice_continuous = False
|
||
_should_cancel_voice = True
|
||
if _should_cancel_voice:
|
||
_cprint(f"\n{_DIM}Recording cancelled.{_RST}")
|
||
threading.Thread(
|
||
target=_recorder_ref.cancel, daemon=True
|
||
).start()
|
||
event.app.invalidate()
|
||
return
|
||
|
||
# Cancel sudo prompt
|
||
if self._sudo_state:
|
||
self._sudo_state["response_queue"].put("")
|
||
self._sudo_state = None
|
||
event.app.current_buffer.reset()
|
||
event.app.invalidate()
|
||
return
|
||
|
||
# Cancel secret prompt
|
||
if self._secret_state:
|
||
self._cancel_secret_capture()
|
||
event.app.current_buffer.reset()
|
||
event.app.invalidate()
|
||
return
|
||
|
||
# Cancel approval prompt (deny)
|
||
if self._approval_state:
|
||
self._approval_state["response_queue"].put("deny")
|
||
self._approval_state = None
|
||
event.app.invalidate()
|
||
return
|
||
|
||
# Cancel clarify prompt
|
||
if self._clarify_state:
|
||
self._clarify_state["response_queue"].put(
|
||
"The user cancelled. Use your best judgement to proceed."
|
||
)
|
||
self._clarify_state = None
|
||
self._clarify_freetext = False
|
||
event.app.current_buffer.reset()
|
||
event.app.invalidate()
|
||
return
|
||
|
||
if self._agent_running and self.agent:
|
||
if now - self._last_ctrl_c_time < 2.0:
|
||
print("\n⚡ Force exiting...")
|
||
self._should_exit = True
|
||
event.app.exit()
|
||
return
|
||
|
||
self._last_ctrl_c_time = now
|
||
print("\n⚡ Interrupting agent... (press Ctrl+C again to force exit)")
|
||
self.agent.interrupt()
|
||
else:
|
||
# If there's text or images, clear them (like bash).
|
||
# If everything is already empty, exit.
|
||
if event.app.current_buffer.text or self._attached_images:
|
||
event.app.current_buffer.reset()
|
||
self._attached_images.clear()
|
||
event.app.invalidate()
|
||
else:
|
||
self._should_exit = True
|
||
event.app.exit()
|
||
|
||
@kb.add('c-d')
|
||
def handle_ctrl_d(event):
|
||
"""Handle Ctrl+D - exit."""
|
||
self._should_exit = True
|
||
event.app.exit()
|
||
|
||
# Voice push-to-talk key: configurable via config.yaml (voice.record_key)
|
||
# Default: Ctrl+B (avoids conflict with Ctrl+R readline reverse-search)
|
||
# Config uses "ctrl+b" format; prompt_toolkit expects "c-b" format.
|
||
try:
|
||
from hermes_cli.config import load_config
|
||
_raw_key = load_config().get("voice", {}).get("record_key", "ctrl+b")
|
||
_voice_key = _raw_key.lower().replace("ctrl+", "c-").replace("alt+", "a-")
|
||
except Exception:
|
||
_voice_key = "c-b"
|
||
|
||
@kb.add(_voice_key)
|
||
def handle_voice_record(event):
|
||
"""Toggle voice recording when voice mode is active.
|
||
|
||
IMPORTANT: This handler runs in prompt_toolkit's event-loop thread.
|
||
Any blocking call here (locks, sd.wait, disk I/O) freezes the
|
||
entire UI. All heavy work is dispatched to daemon threads.
|
||
"""
|
||
if not cli_ref._voice_mode:
|
||
return
|
||
# Always allow STOPPING a recording (even when agent is running)
|
||
if cli_ref._voice_recording:
|
||
# Manual stop via push-to-talk key: stop continuous mode
|
||
with cli_ref._voice_lock:
|
||
cli_ref._voice_continuous = False
|
||
# Flag clearing is handled atomically inside _voice_stop_and_transcribe
|
||
event.app.invalidate()
|
||
threading.Thread(
|
||
target=cli_ref._voice_stop_and_transcribe,
|
||
daemon=True,
|
||
).start()
|
||
else:
|
||
# Guard: don't START recording during agent run or interactive prompts
|
||
if cli_ref._agent_running:
|
||
return
|
||
if cli_ref._clarify_state or cli_ref._sudo_state or cli_ref._approval_state:
|
||
return
|
||
# Guard: don't start while a previous stop/transcribe cycle is
|
||
# still running — recorder.stop() holds AudioRecorder._lock and
|
||
# start() would block the event-loop thread waiting for it.
|
||
if cli_ref._voice_processing:
|
||
return
|
||
|
||
# Interrupt TTS if playing, so user can start talking.
|
||
# stop_playback() is fast (just terminates a subprocess).
|
||
if not cli_ref._voice_tts_done.is_set():
|
||
try:
|
||
from tools.voice_mode import stop_playback
|
||
stop_playback()
|
||
cli_ref._voice_tts_done.set()
|
||
except Exception:
|
||
pass
|
||
|
||
with cli_ref._voice_lock:
|
||
cli_ref._voice_continuous = True
|
||
|
||
# Dispatch to a daemon thread so play_beep(sd.wait),
|
||
# AudioRecorder.start(lock acquire), and config I/O
|
||
# never block the prompt_toolkit event loop.
|
||
def _start_recording():
|
||
try:
|
||
cli_ref._voice_start_recording()
|
||
if hasattr(cli_ref, '_app') and cli_ref._app:
|
||
cli_ref._app.invalidate()
|
||
except Exception as e:
|
||
_cprint(f"\n{_DIM}Voice recording failed: {e}{_RST}")
|
||
|
||
threading.Thread(target=_start_recording, daemon=True).start()
|
||
event.app.invalidate()
|
||
from prompt_toolkit.keys import Keys
|
||
|
||
@kb.add(Keys.BracketedPaste, eager=True)
|
||
def handle_paste(event):
|
||
"""Handle terminal paste — detect clipboard images.
|
||
|
||
When the terminal supports bracketed paste, Ctrl+V / Cmd+V
|
||
triggers this with the pasted text. We also check the
|
||
clipboard for an image on every paste event.
|
||
"""
|
||
pasted_text = event.data or ""
|
||
if self._try_attach_clipboard_image():
|
||
event.app.invalidate()
|
||
if pasted_text:
|
||
event.current_buffer.insert_text(pasted_text)
|
||
|
||
@kb.add('c-v')
|
||
def handle_ctrl_v(event):
|
||
"""Fallback image paste for terminals without bracketed paste.
|
||
|
||
On Linux terminals (GNOME Terminal, Konsole, etc.), Ctrl+V
|
||
sends raw byte 0x16 instead of triggering a paste. This
|
||
binding catches that and checks the clipboard for images.
|
||
On terminals that DO intercept Ctrl+V for paste (macOS
|
||
Terminal, iTerm2, VSCode, Windows Terminal), the bracketed
|
||
paste handler fires instead and this binding never triggers.
|
||
"""
|
||
if self._try_attach_clipboard_image():
|
||
event.app.invalidate()
|
||
|
||
@kb.add('escape', 'v')
|
||
def handle_alt_v(event):
|
||
"""Alt+V — paste image from clipboard.
|
||
|
||
Alt key combos pass through all terminal emulators (sent as
|
||
ESC + key), unlike Ctrl+V which terminals intercept for text
|
||
paste. This is the reliable way to attach clipboard images
|
||
on WSL2, VSCode, and any terminal over SSH where Ctrl+V
|
||
can't reach the application for image-only clipboard.
|
||
"""
|
||
if self._try_attach_clipboard_image():
|
||
event.app.invalidate()
|
||
else:
|
||
# No image found — show a hint
|
||
pass # silent when no image (avoid noise on accidental press)
|
||
|
||
# Dynamic prompt: shows Hermes symbol when agent is working,
|
||
# or answer prompt when clarify freetext mode is active.
|
||
cli_ref = self
|
||
|
||
def get_prompt():
|
||
return cli_ref._get_tui_prompt_fragments()
|
||
|
||
# Create the input area with multiline (shift+enter), autocomplete, and paste handling
|
||
from prompt_toolkit.auto_suggest import AutoSuggestFromHistory
|
||
|
||
def _get_model_completer_info() -> dict:
|
||
"""Return provider/model info for /model autocomplete."""
|
||
try:
|
||
from hermes_cli.models import (
|
||
_PROVIDER_LABELS, _PROVIDER_MODELS, normalize_provider,
|
||
provider_model_ids,
|
||
)
|
||
current = getattr(cli_ref, "provider", None) or getattr(cli_ref, "requested_provider", "openrouter")
|
||
current = normalize_provider(current)
|
||
|
||
# Provider map: id -> label (only providers with known models)
|
||
providers = {}
|
||
for pid, plabel in _PROVIDER_LABELS.items():
|
||
providers[pid] = plabel
|
||
|
||
def models_for(provider_name: str) -> list[str]:
|
||
norm = normalize_provider(provider_name)
|
||
return provider_model_ids(norm)
|
||
|
||
return {
|
||
"current_provider": current,
|
||
"providers": providers,
|
||
"models_for": models_for,
|
||
}
|
||
except Exception:
|
||
return {}
|
||
|
||
_completer = SlashCommandCompleter(
|
||
skill_commands_provider=lambda: _skill_commands,
|
||
model_completer_provider=_get_model_completer_info,
|
||
)
|
||
input_area = TextArea(
|
||
height=Dimension(min=1, max=8, preferred=1),
|
||
prompt=get_prompt,
|
||
style='class:input-area',
|
||
multiline=True,
|
||
wrap_lines=True,
|
||
read_only=Condition(lambda: bool(cli_ref._command_running)),
|
||
history=FileHistory(str(self._history_file)),
|
||
completer=_completer,
|
||
complete_while_typing=True,
|
||
auto_suggest=SlashCommandAutoSuggest(
|
||
history_suggest=AutoSuggestFromHistory(),
|
||
completer=_completer,
|
||
),
|
||
)
|
||
|
||
# Dynamic height: accounts for both explicit newlines AND visual
|
||
# wrapping of long lines so the input area always fits its content.
|
||
def _input_height():
|
||
try:
|
||
doc = input_area.buffer.document
|
||
prompt_width = max(2, len(self._get_tui_prompt_text()))
|
||
available_width = shutil.get_terminal_size().columns - prompt_width
|
||
if available_width < 10:
|
||
available_width = 40
|
||
visual_lines = 0
|
||
for line in doc.lines:
|
||
# Each logical line takes at least 1 visual row; long lines wrap
|
||
if len(line) == 0:
|
||
visual_lines += 1
|
||
else:
|
||
visual_lines += max(1, -(-len(line) // available_width)) # ceil division
|
||
return min(max(visual_lines, 1), 8)
|
||
except Exception:
|
||
return 1
|
||
|
||
input_area.window.height = _input_height
|
||
|
||
# Paste collapsing: detect large pastes and save to temp file
|
||
_paste_counter = [0]
|
||
_prev_text_len = [0]
|
||
|
||
def _on_text_changed(buf):
|
||
"""Detect large pastes and collapse them to a file reference."""
|
||
text = buf.text
|
||
line_count = text.count('\n')
|
||
chars_added = len(text) - _prev_text_len[0]
|
||
_prev_text_len[0] = len(text)
|
||
# Heuristic: a real paste adds many characters at once (not just a
|
||
# single newline from Alt+Enter) AND the result has 5+ lines.
|
||
if line_count >= 5 and chars_added > 1 and not text.startswith('/'):
|
||
_paste_counter[0] += 1
|
||
# Save to temp file
|
||
paste_dir = _hermes_home / "pastes"
|
||
paste_dir.mkdir(parents=True, exist_ok=True)
|
||
paste_file = paste_dir / f"paste_{_paste_counter[0]}_{datetime.now().strftime('%H%M%S')}.txt"
|
||
paste_file.write_text(text, encoding="utf-8")
|
||
# Replace buffer with compact reference
|
||
buf.text = f"[Pasted text #{_paste_counter[0]}: {line_count + 1} lines → {paste_file}]"
|
||
buf.cursor_position = len(buf.text)
|
||
|
||
input_area.buffer.on_text_changed += _on_text_changed
|
||
|
||
# --- Input processors for password masking and inline placeholder ---
|
||
|
||
# Mask input with '*' when the sudo password prompt is active
|
||
input_area.control.input_processors.append(
|
||
ConditionalProcessor(
|
||
PasswordProcessor(),
|
||
filter=Condition(
|
||
lambda: bool(cli_ref._sudo_state) or bool(cli_ref._secret_state)
|
||
),
|
||
)
|
||
)
|
||
|
||
class _PlaceholderProcessor(Processor):
|
||
"""Render grayed-out placeholder text inside the input when empty."""
|
||
def __init__(self, get_text):
|
||
self._get_text = get_text
|
||
|
||
def apply_transformation(self, ti):
|
||
if not ti.document.text and ti.lineno == 0:
|
||
text = self._get_text()
|
||
if text:
|
||
# Append after existing fragments (preserves the ❯ prompt)
|
||
return Transformation(fragments=ti.fragments + [('class:placeholder', text)])
|
||
return Transformation(fragments=ti.fragments)
|
||
|
||
def _get_placeholder():
|
||
if cli_ref._voice_recording:
|
||
return "recording... Ctrl+B to stop, Ctrl+C to cancel"
|
||
if cli_ref._voice_processing:
|
||
return "transcribing..."
|
||
if cli_ref._sudo_state:
|
||
return "type password (hidden), Enter to skip"
|
||
if cli_ref._secret_state:
|
||
return "type secret (hidden), Enter to skip"
|
||
if cli_ref._approval_state:
|
||
return ""
|
||
if cli_ref._clarify_freetext:
|
||
return "type your answer here and press Enter"
|
||
if cli_ref._clarify_state:
|
||
return ""
|
||
if cli_ref._command_running:
|
||
frame = cli_ref._command_spinner_frame()
|
||
status = cli_ref._command_status or "Processing command..."
|
||
return f"{frame} {status}"
|
||
if cli_ref._agent_running:
|
||
return "type a message + Enter to interrupt, Ctrl+C to cancel"
|
||
if cli_ref._voice_mode:
|
||
return "type or Ctrl+B to record"
|
||
return ""
|
||
|
||
input_area.control.input_processors.append(_PlaceholderProcessor(_get_placeholder))
|
||
|
||
# Hint line above input: shown only for interactive prompts that need
|
||
# extra instructions (sudo countdown, approval navigation, clarify).
|
||
# The agent-running interrupt hint is now an inline placeholder above.
|
||
def get_hint_text():
|
||
import time as _time
|
||
|
||
if cli_ref._sudo_state:
|
||
remaining = max(0, int(cli_ref._sudo_deadline - _time.monotonic()))
|
||
return [
|
||
('class:hint', ' password hidden · Enter to skip'),
|
||
('class:clarify-countdown', f' ({remaining}s)'),
|
||
]
|
||
|
||
if cli_ref._secret_state:
|
||
remaining = max(0, int(cli_ref._secret_deadline - _time.monotonic()))
|
||
return [
|
||
('class:hint', ' secret hidden · Enter to skip'),
|
||
('class:clarify-countdown', f' ({remaining}s)'),
|
||
]
|
||
|
||
if cli_ref._approval_state:
|
||
remaining = max(0, int(cli_ref._approval_deadline - _time.monotonic()))
|
||
return [
|
||
('class:hint', ' ↑/↓ to select, Enter to confirm'),
|
||
('class:clarify-countdown', f' ({remaining}s)'),
|
||
]
|
||
|
||
if cli_ref._clarify_state:
|
||
remaining = max(0, int(cli_ref._clarify_deadline - _time.monotonic()))
|
||
countdown = f' ({remaining}s)' if cli_ref._clarify_deadline else ''
|
||
if cli_ref._clarify_freetext:
|
||
return [
|
||
('class:hint', ' type your answer and press Enter'),
|
||
('class:clarify-countdown', countdown),
|
||
]
|
||
return [
|
||
('class:hint', ' ↑/↓ to select, Enter to confirm'),
|
||
('class:clarify-countdown', countdown),
|
||
]
|
||
|
||
if cli_ref._command_running:
|
||
frame = cli_ref._command_spinner_frame()
|
||
return [
|
||
('class:hint', f' {frame} command in progress · input temporarily disabled'),
|
||
]
|
||
|
||
return []
|
||
|
||
def get_hint_height():
|
||
if cli_ref._sudo_state or cli_ref._secret_state or cli_ref._approval_state or cli_ref._clarify_state or cli_ref._command_running:
|
||
return 1
|
||
# Keep a 1-line spacer while agent runs so output doesn't push
|
||
# right up against the top rule of the input area
|
||
return 1 if cli_ref._agent_running else 0
|
||
|
||
def get_spinner_text():
|
||
txt = cli_ref._spinner_text
|
||
if not txt:
|
||
return []
|
||
return [('class:hint', f' {txt}')]
|
||
|
||
def get_spinner_height():
|
||
return 1 if cli_ref._spinner_text else 0
|
||
|
||
spinner_widget = Window(
|
||
content=FormattedTextControl(get_spinner_text),
|
||
height=get_spinner_height,
|
||
)
|
||
|
||
spacer = Window(
|
||
content=FormattedTextControl(get_hint_text),
|
||
height=get_hint_height,
|
||
)
|
||
|
||
# --- Clarify tool: dynamic display widget for questions + choices ---
|
||
|
||
def _panel_box_width(title: str, content_lines: list[str], min_width: int = 46, max_width: int = 76) -> int:
|
||
"""Choose a stable panel width wide enough for the title and content."""
|
||
term_cols = shutil.get_terminal_size((100, 20)).columns
|
||
longest = max([len(title)] + [len(line) for line in content_lines] + [min_width - 4])
|
||
inner = min(max(longest + 4, min_width - 2), max_width - 2, max(24, term_cols - 6))
|
||
return inner + 2 # account for the single leading/trailing spaces inside borders
|
||
|
||
def _wrap_panel_text(text: str, width: int, subsequent_indent: str = "") -> list[str]:
|
||
wrapped = textwrap.wrap(
|
||
text,
|
||
width=max(8, width),
|
||
break_long_words=False,
|
||
break_on_hyphens=False,
|
||
subsequent_indent=subsequent_indent,
|
||
)
|
||
return wrapped or [""]
|
||
|
||
def _append_panel_line(lines, border_style: str, content_style: str, text: str, box_width: int) -> None:
|
||
inner_width = max(0, box_width - 2)
|
||
lines.append((border_style, "│ "))
|
||
lines.append((content_style, text.ljust(inner_width)))
|
||
lines.append((border_style, " │\n"))
|
||
|
||
def _append_blank_panel_line(lines, border_style: str, box_width: int) -> None:
|
||
lines.append((border_style, "│" + (" " * box_width) + "│\n"))
|
||
|
||
def _get_clarify_display():
|
||
"""Build styled text for the clarify question/choices panel."""
|
||
state = cli_ref._clarify_state
|
||
if not state:
|
||
return []
|
||
|
||
question = state["question"]
|
||
choices = state.get("choices") or []
|
||
selected = state.get("selected", 0)
|
||
preview_lines = _wrap_panel_text(question, 60)
|
||
for i, choice in enumerate(choices):
|
||
prefix = "❯ " if i == selected and not cli_ref._clarify_freetext else " "
|
||
preview_lines.extend(_wrap_panel_text(f"{prefix}{choice}", 60, subsequent_indent=" "))
|
||
other_label = (
|
||
"❯ Other (type below)" if cli_ref._clarify_freetext
|
||
else "❯ Other (type your answer)" if selected == len(choices)
|
||
else " Other (type your answer)"
|
||
)
|
||
preview_lines.extend(_wrap_panel_text(other_label, 60, subsequent_indent=" "))
|
||
box_width = _panel_box_width("Hermes needs your input", preview_lines)
|
||
inner_text_width = max(8, box_width - 2)
|
||
|
||
lines = []
|
||
# Box top border
|
||
lines.append(('class:clarify-border', '╭─ '))
|
||
lines.append(('class:clarify-title', 'Hermes needs your input'))
|
||
lines.append(('class:clarify-border', ' ' + ('─' * max(0, box_width - len("Hermes needs your input") - 3)) + '╮\n'))
|
||
_append_blank_panel_line(lines, 'class:clarify-border', box_width)
|
||
|
||
# Question text
|
||
for wrapped in _wrap_panel_text(question, inner_text_width):
|
||
_append_panel_line(lines, 'class:clarify-border', 'class:clarify-question', wrapped, box_width)
|
||
_append_blank_panel_line(lines, 'class:clarify-border', box_width)
|
||
|
||
if cli_ref._clarify_freetext and not choices:
|
||
guidance = "Type your answer in the prompt below, then press Enter."
|
||
for wrapped in _wrap_panel_text(guidance, inner_text_width):
|
||
_append_panel_line(lines, 'class:clarify-border', 'class:clarify-choice', wrapped, box_width)
|
||
_append_blank_panel_line(lines, 'class:clarify-border', box_width)
|
||
|
||
if choices:
|
||
# Multiple-choice mode: show selectable options
|
||
for i, choice in enumerate(choices):
|
||
style = 'class:clarify-selected' if i == selected and not cli_ref._clarify_freetext else 'class:clarify-choice'
|
||
prefix = '❯ ' if i == selected and not cli_ref._clarify_freetext else ' '
|
||
wrapped_lines = _wrap_panel_text(f"{prefix}{choice}", inner_text_width, subsequent_indent=" ")
|
||
for wrapped in wrapped_lines:
|
||
_append_panel_line(lines, 'class:clarify-border', style, wrapped, box_width)
|
||
|
||
# "Other" option (5th line, only shown when choices exist)
|
||
other_idx = len(choices)
|
||
if selected == other_idx and not cli_ref._clarify_freetext:
|
||
other_style = 'class:clarify-selected'
|
||
other_label = '❯ Other (type your answer)'
|
||
elif cli_ref._clarify_freetext:
|
||
other_style = 'class:clarify-active-other'
|
||
other_label = '❯ Other (type below)'
|
||
else:
|
||
other_style = 'class:clarify-choice'
|
||
other_label = ' Other (type your answer)'
|
||
for wrapped in _wrap_panel_text(other_label, inner_text_width, subsequent_indent=" "):
|
||
_append_panel_line(lines, 'class:clarify-border', other_style, wrapped, box_width)
|
||
|
||
_append_blank_panel_line(lines, 'class:clarify-border', box_width)
|
||
lines.append(('class:clarify-border', '╰' + ('─' * box_width) + '╯\n'))
|
||
return lines
|
||
|
||
clarify_widget = ConditionalContainer(
|
||
Window(
|
||
FormattedTextControl(_get_clarify_display),
|
||
wrap_lines=True,
|
||
),
|
||
filter=Condition(lambda: cli_ref._clarify_state is not None),
|
||
)
|
||
|
||
# --- Sudo password: display widget ---
|
||
|
||
def _get_sudo_display():
|
||
state = cli_ref._sudo_state
|
||
if not state:
|
||
return []
|
||
title = '🔐 Sudo Password Required'
|
||
body = 'Enter password below (hidden), or press Enter to skip'
|
||
box_width = _panel_box_width(title, [body])
|
||
inner = max(0, box_width - 2)
|
||
lines = []
|
||
lines.append(('class:sudo-border', '╭─ '))
|
||
lines.append(('class:sudo-title', title))
|
||
lines.append(('class:sudo-border', ' ' + ('─' * max(0, box_width - len(title) - 3)) + '╮\n'))
|
||
_append_blank_panel_line(lines, 'class:sudo-border', box_width)
|
||
_append_panel_line(lines, 'class:sudo-border', 'class:sudo-text', body, box_width)
|
||
_append_blank_panel_line(lines, 'class:sudo-border', box_width)
|
||
lines.append(('class:sudo-border', '╰' + ('─' * box_width) + '╯\n'))
|
||
return lines
|
||
|
||
sudo_widget = ConditionalContainer(
|
||
Window(
|
||
FormattedTextControl(_get_sudo_display),
|
||
wrap_lines=True,
|
||
),
|
||
filter=Condition(lambda: cli_ref._sudo_state is not None),
|
||
)
|
||
|
||
def _get_secret_display():
|
||
state = cli_ref._secret_state
|
||
if not state:
|
||
return []
|
||
|
||
title = '🔑 Skill Setup Required'
|
||
prompt = state.get("prompt") or f"Enter value for {state.get('var_name', 'secret')}"
|
||
metadata = state.get("metadata") or {}
|
||
help_text = metadata.get("help")
|
||
body = 'Enter secret below (hidden), or press Enter to skip'
|
||
content_lines = [prompt, body]
|
||
if help_text:
|
||
content_lines.insert(1, str(help_text))
|
||
box_width = _panel_box_width(title, content_lines)
|
||
lines = []
|
||
lines.append(('class:sudo-border', '╭─ '))
|
||
lines.append(('class:sudo-title', title))
|
||
lines.append(('class:sudo-border', ' ' + ('─' * max(0, box_width - len(title) - 3)) + '╮\n'))
|
||
_append_blank_panel_line(lines, 'class:sudo-border', box_width)
|
||
_append_panel_line(lines, 'class:sudo-border', 'class:sudo-text', prompt, box_width)
|
||
if help_text:
|
||
_append_panel_line(lines, 'class:sudo-border', 'class:sudo-text', str(help_text), box_width)
|
||
_append_blank_panel_line(lines, 'class:sudo-border', box_width)
|
||
_append_panel_line(lines, 'class:sudo-border', 'class:sudo-text', body, box_width)
|
||
_append_blank_panel_line(lines, 'class:sudo-border', box_width)
|
||
lines.append(('class:sudo-border', '╰' + ('─' * box_width) + '╯\n'))
|
||
return lines
|
||
|
||
secret_widget = ConditionalContainer(
|
||
Window(
|
||
FormattedTextControl(_get_secret_display),
|
||
wrap_lines=True,
|
||
),
|
||
filter=Condition(lambda: cli_ref._secret_state is not None),
|
||
)
|
||
|
||
# --- Dangerous command approval: display widget ---
|
||
|
||
def _get_approval_display():
|
||
return cli_ref._get_approval_display_fragments()
|
||
|
||
approval_widget = ConditionalContainer(
|
||
Window(
|
||
FormattedTextControl(_get_approval_display),
|
||
wrap_lines=True,
|
||
),
|
||
filter=Condition(lambda: cli_ref._approval_state is not None),
|
||
)
|
||
|
||
# Horizontal rules above and below the input (bronze, 1 line each).
|
||
# The bottom rule moves down as the TextArea grows with newlines.
|
||
# Using char='─' instead of hardcoded repetition so the rule
|
||
# always spans the full terminal width on any screen size.
|
||
input_rule_top = Window(
|
||
char='─',
|
||
height=1,
|
||
style='class:input-rule',
|
||
)
|
||
input_rule_bot = Window(
|
||
char='─',
|
||
height=1,
|
||
style='class:input-rule',
|
||
)
|
||
|
||
# Image attachment indicator — shows badges like [📎 Image #1] above input
|
||
cli_ref = self
|
||
|
||
def _get_image_bar():
|
||
if not cli_ref._attached_images:
|
||
return []
|
||
base = cli_ref._image_counter - len(cli_ref._attached_images) + 1
|
||
badges = " ".join(
|
||
f"[📎 Image #{base + i}]"
|
||
for i in range(len(cli_ref._attached_images))
|
||
)
|
||
return [("class:image-badge", f" {badges} ")]
|
||
|
||
image_bar = Window(
|
||
content=FormattedTextControl(_get_image_bar),
|
||
height=Condition(lambda: bool(cli_ref._attached_images)),
|
||
)
|
||
|
||
# Persistent voice mode status bar (visible only when voice mode is on)
|
||
def _get_voice_status():
|
||
if cli_ref._voice_recording:
|
||
return [('class:voice-status-recording', ' ● REC Ctrl+B to stop ')]
|
||
if cli_ref._voice_processing:
|
||
return [('class:voice-status', ' ◉ Transcribing... ')]
|
||
tts = " | TTS on" if cli_ref._voice_tts else ""
|
||
cont = " | Continuous" if cli_ref._voice_continuous else ""
|
||
return [('class:voice-status', f' 🎤 Voice mode{tts}{cont} — Ctrl+B to record ')]
|
||
|
||
voice_status_bar = ConditionalContainer(
|
||
Window(
|
||
FormattedTextControl(_get_voice_status),
|
||
height=1,
|
||
),
|
||
filter=Condition(lambda: cli_ref._voice_mode),
|
||
)
|
||
|
||
status_bar = ConditionalContainer(
|
||
Window(
|
||
content=FormattedTextControl(lambda: cli_ref._get_status_bar_fragments()),
|
||
height=1,
|
||
),
|
||
filter=Condition(lambda: cli_ref._status_bar_visible),
|
||
)
|
||
|
||
# Layout: interactive prompt widgets + ruled input at bottom.
|
||
# The sudo, approval, and clarify widgets appear above the input when
|
||
# the corresponding interactive prompt is active.
|
||
layout = Layout(
|
||
HSplit([
|
||
Window(height=0),
|
||
sudo_widget,
|
||
secret_widget,
|
||
approval_widget,
|
||
clarify_widget,
|
||
spinner_widget,
|
||
spacer,
|
||
status_bar,
|
||
input_rule_top,
|
||
image_bar,
|
||
input_area,
|
||
input_rule_bot,
|
||
voice_status_bar,
|
||
CompletionsMenu(max_height=12, scroll_offset=1),
|
||
])
|
||
)
|
||
|
||
# Style for the application
|
||
self._tui_style_base = {
|
||
'input-area': '#FFF8DC',
|
||
'placeholder': '#555555 italic',
|
||
'prompt': '#FFF8DC',
|
||
'prompt-working': '#888888 italic',
|
||
'hint': '#555555 italic',
|
||
'status-bar': 'bg:#1a1a2e #C0C0C0',
|
||
'status-bar-strong': 'bg:#1a1a2e #FFD700 bold',
|
||
'status-bar-dim': 'bg:#1a1a2e #8B8682',
|
||
'status-bar-good': 'bg:#1a1a2e #8FBC8F bold',
|
||
'status-bar-warn': 'bg:#1a1a2e #FFD700 bold',
|
||
'status-bar-bad': 'bg:#1a1a2e #FF8C00 bold',
|
||
'status-bar-critical': 'bg:#1a1a2e #FF6B6B bold',
|
||
# Bronze horizontal rules around the input area
|
||
'input-rule': '#CD7F32',
|
||
# Clipboard image attachment badges
|
||
'image-badge': '#87CEEB bold',
|
||
'completion-menu': 'bg:#1a1a2e #FFF8DC',
|
||
'completion-menu.completion': 'bg:#1a1a2e #FFF8DC',
|
||
'completion-menu.completion.current': 'bg:#333355 #FFD700',
|
||
'completion-menu.meta.completion': 'bg:#1a1a2e #888888',
|
||
'completion-menu.meta.completion.current': 'bg:#333355 #FFBF00',
|
||
# Clarify question panel
|
||
'clarify-border': '#CD7F32',
|
||
'clarify-title': '#FFD700 bold',
|
||
'clarify-question': '#FFF8DC bold',
|
||
'clarify-choice': '#AAAAAA',
|
||
'clarify-selected': '#FFD700 bold',
|
||
'clarify-active-other': '#FFD700 italic',
|
||
'clarify-countdown': '#CD7F32',
|
||
# Sudo password panel
|
||
'sudo-prompt': '#FF6B6B bold',
|
||
'sudo-border': '#CD7F32',
|
||
'sudo-title': '#FF6B6B bold',
|
||
'sudo-text': '#FFF8DC',
|
||
# Dangerous command approval panel
|
||
'approval-border': '#CD7F32',
|
||
'approval-title': '#FF8C00 bold',
|
||
'approval-desc': '#FFF8DC bold',
|
||
'approval-cmd': '#AAAAAA italic',
|
||
'approval-choice': '#AAAAAA',
|
||
'approval-selected': '#FFD700 bold',
|
||
# Voice mode
|
||
'voice-prompt': '#87CEEB',
|
||
'voice-recording': '#FF4444 bold',
|
||
'voice-processing': '#FFA500 italic',
|
||
'voice-status': 'bg:#1a1a2e #87CEEB',
|
||
'voice-status-recording': 'bg:#1a1a2e #FF4444 bold',
|
||
}
|
||
style = PTStyle.from_dict(self._build_tui_style_dict())
|
||
|
||
# Create the application
|
||
app = Application(
|
||
layout=layout,
|
||
key_bindings=kb,
|
||
style=style,
|
||
full_screen=False,
|
||
mouse_support=False,
|
||
**({'cursor': _STEADY_CURSOR} if _STEADY_CURSOR is not None else {}),
|
||
)
|
||
self._app = app # Store reference for clarify_callback
|
||
|
||
def spinner_loop():
|
||
import time as _time
|
||
|
||
last_idle_refresh = 0.0
|
||
while not self._should_exit:
|
||
if not self._app:
|
||
_time.sleep(0.1)
|
||
continue
|
||
if self._command_running:
|
||
self._invalidate(min_interval=0.1)
|
||
_time.sleep(0.1)
|
||
else:
|
||
now = _time.monotonic()
|
||
if now - last_idle_refresh >= 1.0:
|
||
last_idle_refresh = now
|
||
self._invalidate(min_interval=1.0)
|
||
_time.sleep(0.2)
|
||
|
||
spinner_thread = threading.Thread(target=spinner_loop, daemon=True)
|
||
spinner_thread.start()
|
||
|
||
# Background thread to process inputs and run agent
|
||
def process_loop():
|
||
while not self._should_exit:
|
||
try:
|
||
# Check for pending input with timeout
|
||
try:
|
||
user_input = self._pending_input.get(timeout=0.1)
|
||
except queue.Empty:
|
||
# Periodic config watcher — auto-reload MCP on mcp_servers change
|
||
if not self._agent_running:
|
||
self._check_config_mcp_changes()
|
||
continue
|
||
|
||
if not user_input:
|
||
continue
|
||
|
||
# Unpack image payload: (text, [Path, ...]) or plain str
|
||
submit_images = []
|
||
if isinstance(user_input, tuple):
|
||
user_input, submit_images = user_input
|
||
|
||
# Check for commands
|
||
if isinstance(user_input, str) and user_input.startswith("/"):
|
||
_cprint(f"\n⚙️ {user_input}")
|
||
if not self.process_command(user_input):
|
||
self._should_exit = True
|
||
# Schedule app exit
|
||
if app.is_running:
|
||
app.exit()
|
||
continue
|
||
|
||
# Expand paste references back to full content
|
||
import re as _re
|
||
paste_match = _re.match(r'\[Pasted text #\d+: \d+ lines → (.+)\]', user_input) if isinstance(user_input, str) else None
|
||
if paste_match:
|
||
paste_path = Path(paste_match.group(1))
|
||
if paste_path.exists():
|
||
full_text = paste_path.read_text(encoding="utf-8")
|
||
line_count = full_text.count('\n') + 1
|
||
print()
|
||
ChatConsole().print(
|
||
f"[bold {_accent_hex()}]●[/] [bold]{_escape(f'[Pasted text: {line_count} lines]')}[/]"
|
||
)
|
||
user_input = full_text
|
||
else:
|
||
print()
|
||
ChatConsole().print(f"[bold {_accent_hex()}]●[/] [bold]{_escape(user_input)}[/]")
|
||
else:
|
||
if '\n' in user_input:
|
||
first_line = user_input.split('\n')[0]
|
||
line_count = user_input.count('\n') + 1
|
||
print()
|
||
ChatConsole().print(
|
||
f"[bold {_accent_hex()}]●[/] [bold]{_escape(first_line)}[/] "
|
||
f"[dim](+{line_count - 1} lines)[/]"
|
||
)
|
||
else:
|
||
print()
|
||
ChatConsole().print(f"[bold {_accent_hex()}]●[/] [bold]{_escape(user_input)}[/]")
|
||
|
||
# Show image attachment count
|
||
if submit_images:
|
||
n = len(submit_images)
|
||
_cprint(f" {_DIM}📎 {n} image{'s' if n > 1 else ''} attached{_RST}")
|
||
|
||
# Regular chat - run agent
|
||
self._agent_running = True
|
||
app.invalidate() # Refresh status line
|
||
|
||
try:
|
||
self.chat(user_input, images=submit_images or None)
|
||
finally:
|
||
self._agent_running = False
|
||
self._spinner_text = ""
|
||
app.invalidate() # Refresh status line
|
||
|
||
# Continuous voice: auto-restart recording after agent responds.
|
||
# Dispatch to a daemon thread so play_beep (sd.wait) and
|
||
# AudioRecorder.start (lock acquire) never block process_loop —
|
||
# otherwise queued user input would stall silently.
|
||
if self._voice_mode and self._voice_continuous and not self._voice_recording:
|
||
def _restart_recording():
|
||
try:
|
||
if self._voice_tts:
|
||
self._voice_tts_done.wait(timeout=60)
|
||
time.sleep(0.3)
|
||
self._voice_start_recording()
|
||
app.invalidate()
|
||
except Exception as e:
|
||
_cprint(f"{_DIM}Voice auto-restart failed: {e}{_RST}")
|
||
threading.Thread(target=_restart_recording, daemon=True).start()
|
||
|
||
except Exception as e:
|
||
print(f"Error: {e}")
|
||
|
||
# Start processing thread
|
||
process_thread = threading.Thread(target=process_loop, daemon=True)
|
||
process_thread.start()
|
||
|
||
# Register atexit cleanup so resources are freed even on unexpected exit
|
||
atexit.register(_run_cleanup)
|
||
|
||
# Run the application with patch_stdout for proper output handling
|
||
try:
|
||
with patch_stdout():
|
||
app.run()
|
||
except (EOFError, KeyboardInterrupt):
|
||
pass
|
||
finally:
|
||
self._should_exit = True
|
||
# Flush memories before exit (only for substantial conversations)
|
||
if self.agent and self.conversation_history:
|
||
try:
|
||
self.agent.flush_memories(self.conversation_history)
|
||
except Exception:
|
||
pass
|
||
# Shut down voice recorder (release persistent audio stream)
|
||
if hasattr(self, '_voice_recorder') and self._voice_recorder:
|
||
try:
|
||
self._voice_recorder.shutdown()
|
||
except Exception:
|
||
pass
|
||
self._voice_recorder = None
|
||
# Clean up old temp voice recordings
|
||
try:
|
||
from tools.voice_mode import cleanup_temp_recordings
|
||
cleanup_temp_recordings()
|
||
except Exception:
|
||
pass
|
||
# Unregister callbacks to avoid dangling references
|
||
set_sudo_password_callback(None)
|
||
set_approval_callback(None)
|
||
set_secret_capture_callback(None)
|
||
# Flush + shut down Honcho async writer (drains queue before exit)
|
||
if self.agent and getattr(self.agent, '_honcho', None):
|
||
try:
|
||
self.agent._honcho.shutdown()
|
||
except Exception:
|
||
pass
|
||
# Close session in SQLite
|
||
if hasattr(self, '_session_db') and self._session_db and self.agent:
|
||
try:
|
||
self._session_db.end_session(self.agent.session_id, "cli_close")
|
||
except Exception as e:
|
||
logger.debug("Could not close session in DB: %s", e)
|
||
_run_cleanup()
|
||
self._print_exit_summary()
|
||
|
||
|
||
# ============================================================================
|
||
# Main Entry Point
|
||
# ============================================================================
|
||
|
||
def main(
|
||
query: str = None,
|
||
q: str = None,
|
||
toolsets: str = None,
|
||
skills: str | list[str] | tuple[str, ...] = None,
|
||
model: str = None,
|
||
provider: str = None,
|
||
api_key: str = None,
|
||
base_url: str = None,
|
||
max_turns: int = None,
|
||
verbose: bool = False,
|
||
quiet: bool = False,
|
||
compact: bool = False,
|
||
list_tools: bool = False,
|
||
list_toolsets: bool = False,
|
||
gateway: bool = False,
|
||
resume: str = None,
|
||
worktree: bool = False,
|
||
w: bool = False,
|
||
checkpoints: bool = False,
|
||
pass_session_id: bool = False,
|
||
):
|
||
"""
|
||
Hermes Agent CLI - Interactive AI Assistant
|
||
|
||
Args:
|
||
query: Single query to execute (then exit). Alias: -q
|
||
q: Shorthand for --query
|
||
toolsets: Comma-separated list of toolsets to enable (e.g., "web,terminal")
|
||
skills: Comma-separated or repeated list of skills to preload for the session
|
||
model: Model to use (default: anthropic/claude-opus-4-20250514)
|
||
provider: Inference provider ("auto", "openrouter", "nous", "openai-codex", "zai", "kimi-coding", "minimax", "minimax-cn")
|
||
api_key: API key for authentication
|
||
base_url: Base URL for the API
|
||
max_turns: Maximum tool-calling iterations (default: 60)
|
||
verbose: Enable verbose logging
|
||
compact: Use compact display mode
|
||
list_tools: List available tools and exit
|
||
list_toolsets: List available toolsets and exit
|
||
resume: Resume a previous session by its ID (e.g., 20260225_143052_a1b2c3)
|
||
worktree: Run in an isolated git worktree (for parallel agents). Alias: -w
|
||
w: Shorthand for --worktree
|
||
|
||
Examples:
|
||
python cli.py # Start interactive mode
|
||
python cli.py --toolsets web,terminal # Use specific toolsets
|
||
python cli.py --skills hermes-agent-dev,github-auth
|
||
python cli.py -q "What is Python?" # Single query mode
|
||
python cli.py --list-tools # List tools and exit
|
||
python cli.py --resume 20260225_143052_a1b2c3 # Resume session
|
||
python cli.py -w # Start in isolated git worktree
|
||
python cli.py -w -q "Fix issue #123" # Single query in worktree
|
||
"""
|
||
global _active_worktree
|
||
|
||
# Signal to terminal_tool that we're in interactive mode
|
||
# This enables interactive sudo password prompts with timeout
|
||
os.environ["HERMES_INTERACTIVE"] = "1"
|
||
|
||
# Handle gateway mode (messaging + cron)
|
||
if gateway:
|
||
import asyncio
|
||
from gateway.run import start_gateway
|
||
print("Starting Hermes Gateway (messaging platforms)...")
|
||
asyncio.run(start_gateway())
|
||
return
|
||
|
||
# Skip worktree for list commands (they exit immediately)
|
||
if not list_tools and not list_toolsets:
|
||
# ── Git worktree isolation (#652) ──
|
||
# Create an isolated worktree so this agent instance doesn't collide
|
||
# with other agents working on the same repo.
|
||
use_worktree = worktree or w or CLI_CONFIG.get("worktree", False)
|
||
wt_info = None
|
||
if use_worktree:
|
||
# Prune stale worktrees from crashed/killed sessions
|
||
_repo = _git_repo_root()
|
||
if _repo:
|
||
_prune_stale_worktrees(_repo)
|
||
wt_info = _setup_worktree()
|
||
if wt_info:
|
||
_active_worktree = wt_info
|
||
os.environ["TERMINAL_CWD"] = wt_info["path"]
|
||
atexit.register(_cleanup_worktree, wt_info)
|
||
else:
|
||
# Worktree was explicitly requested but setup failed —
|
||
# don't silently run without isolation.
|
||
return
|
||
else:
|
||
wt_info = None
|
||
|
||
# Handle query shorthand
|
||
query = query or q
|
||
|
||
# Parse toolsets - handle both string and tuple/list inputs
|
||
# Default to hermes-cli toolset which includes cronjob management tools
|
||
toolsets_list = None
|
||
if toolsets:
|
||
if isinstance(toolsets, str):
|
||
toolsets_list = [t.strip() for t in toolsets.split(",")]
|
||
elif isinstance(toolsets, (list, tuple)):
|
||
# Fire may pass multiple --toolsets as a tuple
|
||
toolsets_list = []
|
||
for t in toolsets:
|
||
if isinstance(t, str):
|
||
toolsets_list.extend([x.strip() for x in t.split(",")])
|
||
else:
|
||
toolsets_list.append(str(t))
|
||
else:
|
||
# Check config for CLI toolsets, fallback to hermes-cli
|
||
config_cli_toolsets = CLI_CONFIG.get("platform_toolsets", {}).get("cli")
|
||
if config_cli_toolsets and isinstance(config_cli_toolsets, list):
|
||
toolsets_list = config_cli_toolsets
|
||
else:
|
||
toolsets_list = ["hermes-cli"]
|
||
|
||
parsed_skills = _parse_skills_argument(skills)
|
||
|
||
# Create CLI instance
|
||
cli = HermesCLI(
|
||
model=model,
|
||
toolsets=toolsets_list,
|
||
provider=provider,
|
||
api_key=api_key,
|
||
base_url=base_url,
|
||
max_turns=max_turns,
|
||
verbose=verbose,
|
||
compact=compact,
|
||
resume=resume,
|
||
checkpoints=checkpoints,
|
||
pass_session_id=pass_session_id,
|
||
)
|
||
|
||
if parsed_skills:
|
||
skills_prompt, loaded_skills, missing_skills = build_preloaded_skills_prompt(
|
||
parsed_skills,
|
||
task_id=cli.session_id,
|
||
)
|
||
if missing_skills:
|
||
missing_display = ", ".join(missing_skills)
|
||
raise ValueError(f"Unknown skill(s): {missing_display}")
|
||
if skills_prompt:
|
||
cli.system_prompt = "\n\n".join(
|
||
part for part in (cli.system_prompt, skills_prompt) if part
|
||
).strip()
|
||
cli.preloaded_skills = loaded_skills
|
||
|
||
# Inject worktree context into agent's system prompt
|
||
if wt_info:
|
||
wt_note = (
|
||
f"\n\n[System note: You are working in an isolated git worktree at "
|
||
f"{wt_info['path']}. Your branch is `{wt_info['branch']}`. "
|
||
f"Changes here do not affect the main working tree or other agents. "
|
||
f"Remember to commit and push your changes, and create a PR if appropriate. "
|
||
f"The original repo is at {wt_info['repo_root']}.]"
|
||
)
|
||
cli.system_prompt = (cli.system_prompt or "") + wt_note
|
||
|
||
# Handle list commands (don't init agent for these)
|
||
if list_tools:
|
||
cli.show_banner()
|
||
cli.show_tools()
|
||
sys.exit(0)
|
||
|
||
if list_toolsets:
|
||
cli.show_banner()
|
||
cli.show_toolsets()
|
||
sys.exit(0)
|
||
|
||
# Register cleanup for single-query mode (interactive mode registers in run())
|
||
atexit.register(_run_cleanup)
|
||
|
||
# Handle single query mode
|
||
if query:
|
||
if quiet:
|
||
# Quiet mode: suppress banner, spinner, tool previews.
|
||
# Only print the final response and parseable session info.
|
||
cli.tool_progress_mode = "off"
|
||
if cli._ensure_runtime_credentials():
|
||
turn_route = cli._resolve_turn_agent_config(query)
|
||
if turn_route["signature"] != cli._active_agent_route_signature:
|
||
cli.agent = None
|
||
if cli._init_agent(
|
||
model_override=turn_route["model"],
|
||
runtime_override=turn_route["runtime"],
|
||
route_label=turn_route["label"],
|
||
):
|
||
cli.agent.quiet_mode = True
|
||
result = cli.agent.run_conversation(query)
|
||
response = result.get("final_response", "") if isinstance(result, dict) else str(result)
|
||
if response:
|
||
print(response)
|
||
print(f"\nsession_id: {cli.session_id}")
|
||
else:
|
||
cli.show_banner()
|
||
cli.console.print(f"[bold blue]Query:[/] {query}")
|
||
cli.chat(query)
|
||
cli._print_exit_summary()
|
||
return
|
||
|
||
# Run interactive mode
|
||
cli.run()
|
||
|
||
|
||
if __name__ == "__main__":
|
||
fire.Fire(main)
|