"""Optimized dashboard app with improved async handling and non-blocking startup. Key improvements: 1. Background tasks use asyncio.create_task() to avoid blocking startup 2. Chat integrations start in background 3. All startup operations complete quickly 4. Security and logging handled by dedicated middleware """ import asyncio import json import logging import re from contextlib import asynccontextmanager from pathlib import Path from fastapi import FastAPI, Request, WebSocket from fastapi.middleware.cors import CORSMiddleware from fastapi.middleware.trustedhost import TrustedHostMiddleware from fastapi.responses import HTMLResponse from fastapi.staticfiles import StaticFiles from config import settings # Import dedicated middleware from dashboard.middleware.csrf import CSRFMiddleware from dashboard.middleware.rate_limit import RateLimitMiddleware from dashboard.middleware.request_logging import RequestLoggingMiddleware from dashboard.middleware.security_headers import SecurityHeadersMiddleware from dashboard.routes.agents import router as agents_router from dashboard.routes.briefing import router as briefing_router from dashboard.routes.calm import router as calm_router from dashboard.routes.chat_api import router as chat_api_router from dashboard.routes.chat_api_v1 import router as chat_api_v1_router from dashboard.routes.daily_run import router as daily_run_router from dashboard.routes.db_explorer import router as db_explorer_router from dashboard.routes.discord import router as discord_router from dashboard.routes.experiments import router as experiments_router from dashboard.routes.grok import router as grok_router from dashboard.routes.health import router as health_router from dashboard.routes.loop_qa import router as loop_qa_router from dashboard.routes.memory import router as memory_router from dashboard.routes.mobile import router as mobile_router from dashboard.routes.models import api_router as models_api_router from dashboard.routes.models import router as models_router from dashboard.routes.quests import router as quests_router from dashboard.routes.scorecards import router as scorecards_router from dashboard.routes.sovereignty_metrics import router as sovereignty_metrics_router from dashboard.routes.spark import router as spark_router from dashboard.routes.system import router as system_router from dashboard.routes.tasks import router as tasks_router from dashboard.routes.telegram import router as telegram_router from dashboard.routes.thinking import router as thinking_router from dashboard.routes.tools import router as tools_router from dashboard.routes.tower import router as tower_router from dashboard.routes.voice import router as voice_router from dashboard.routes.work_orders import router as work_orders_router from dashboard.routes.world import matrix_router from dashboard.routes.world import router as world_router from timmy.workshop_state import PRESENCE_FILE class _ColorFormatter(logging.Formatter): """ANSI color formatter — red is reserved for ERROR/CRITICAL only.""" RESET = "\033[0m" COLORS = { logging.DEBUG: "\033[37m", # white/gray logging.INFO: "\033[32m", # green logging.WARNING: "\033[33m", # yellow logging.ERROR: "\033[31m", # red logging.CRITICAL: "\033[1;31m", # bold red } def format(self, record: logging.LogRecord) -> str: color = self.COLORS.get(record.levelno, self.RESET) formatted = super().format(record) return f"{color}{formatted}{self.RESET}" def _configure_logging() -> None: """Configure logging with console and optional rotating file handler.""" root_logger = logging.getLogger() root_logger.setLevel(logging.INFO) console = logging.StreamHandler() console.setLevel(logging.INFO) console.setFormatter( _ColorFormatter( "%(asctime)s %(levelname)-8s %(name)s — %(message)s", datefmt="%H:%M:%S", ) ) root_logger.addHandler(console) # Override uvicorn's default colored formatter so all console output # uses our color scheme (red = ERROR/CRITICAL only). for name in ("uvicorn", "uvicorn.error", "uvicorn.access"): uv_logger = logging.getLogger(name) uv_logger.handlers.clear() uv_logger.propagate = True if settings.error_log_enabled: from logging.handlers import RotatingFileHandler log_dir = Path(settings.repo_root) / settings.error_log_dir log_dir.mkdir(parents=True, exist_ok=True) error_file = log_dir / "errors.log" file_handler = RotatingFileHandler( error_file, maxBytes=settings.error_log_max_bytes, backupCount=settings.error_log_backup_count, ) file_handler.setLevel(logging.ERROR) file_handler.setFormatter( logging.Formatter( "%(asctime)s %(levelname)-8s %(name)s — %(message)s\n" " File: %(pathname)s:%(lineno)d\n" " Function: %(funcName)s", datefmt="%Y-%m-%d %H:%M:%S", ) ) root_logger.addHandler(file_handler) _configure_logging() logger = logging.getLogger(__name__) BASE_DIR = Path(__file__).parent PROJECT_ROOT = BASE_DIR.parent.parent _BRIEFING_INTERVAL_HOURS = 6 async def _briefing_scheduler() -> None: """Background task: regenerate Timmy's briefing every 6 hours.""" from infrastructure.notifications.push import notify_briefing_ready from timmy.briefing import engine as briefing_engine await asyncio.sleep(2) while True: try: if briefing_engine.needs_refresh(): logger.info("Generating morning briefing…") briefing = briefing_engine.generate() await notify_briefing_ready(briefing) else: logger.info("Briefing is fresh; skipping generation.") except Exception as exc: logger.error("Briefing scheduler error: %s", exc) await asyncio.sleep(_BRIEFING_INTERVAL_HOURS * 3600) async def _thinking_scheduler() -> None: """Background task: execute Timmy's thinking cycle every N seconds.""" from timmy.thinking import thinking_engine await asyncio.sleep(5) # Stagger after briefing scheduler while True: try: if settings.thinking_enabled: await asyncio.wait_for( thinking_engine.think_once(), timeout=settings.thinking_timeout_seconds, ) except TimeoutError: logger.warning( "Thinking cycle timed out after %ds — Ollama may be unresponsive", settings.thinking_timeout_seconds, ) except asyncio.CancelledError: raise except Exception as exc: logger.error("Thinking scheduler error: %s", exc) await asyncio.sleep(settings.thinking_interval_seconds) async def _loop_qa_scheduler() -> None: """Background task: run capability self-tests on a separate timer. Independent of the thinking loop — runs every N thinking ticks to probe subsystems and detect degradation. """ from timmy.loop_qa import loop_qa_orchestrator await asyncio.sleep(10) # Stagger after thinking scheduler while True: try: if settings.loop_qa_enabled: result = await asyncio.wait_for( loop_qa_orchestrator.run_next_test(), timeout=settings.thinking_timeout_seconds, ) if result: status = "PASS" if result["success"] else "FAIL" logger.info( "Loop QA [%s]: %s — %s", result["capability"], status, result.get("details", "")[:80], ) except TimeoutError: logger.warning( "Loop QA test timed out after %ds", settings.thinking_timeout_seconds, ) except asyncio.CancelledError: raise except Exception as exc: logger.error("Loop QA scheduler error: %s", exc) interval = settings.thinking_interval_seconds * settings.loop_qa_interval_ticks await asyncio.sleep(interval) _PRESENCE_POLL_SECONDS = 30 _PRESENCE_INITIAL_DELAY = 3 _SYNTHESIZED_STATE: dict = { "version": 1, "liveness": None, "current_focus": "", "mood": "idle", "active_threads": [], "recent_events": [], "concerns": [], } async def _presence_watcher() -> None: """Background task: watch ~/.timmy/presence.json and broadcast changes via WS. Polls the file every 30 seconds (matching Timmy's write cadence). If the file doesn't exist, broadcasts a synthesised idle state. """ from infrastructure.ws_manager.handler import ws_manager as ws_mgr await asyncio.sleep(_PRESENCE_INITIAL_DELAY) # Stagger after other schedulers last_mtime: float = 0.0 while True: try: if PRESENCE_FILE.exists(): mtime = PRESENCE_FILE.stat().st_mtime if mtime != last_mtime: last_mtime = mtime raw = await asyncio.to_thread(PRESENCE_FILE.read_text) state = json.loads(raw) await ws_mgr.broadcast("timmy_state", state) else: # File absent — broadcast synthesised state once per cycle if last_mtime != -1.0: last_mtime = -1.0 await ws_mgr.broadcast("timmy_state", _SYNTHESIZED_STATE) except json.JSONDecodeError as exc: logger.warning("presence.json parse error: %s", exc) except Exception as exc: logger.warning("Presence watcher error: %s", exc) await asyncio.sleep(_PRESENCE_POLL_SECONDS) async def _start_chat_integrations_background() -> None: """Background task: start chat integrations without blocking startup.""" from integrations.chat_bridge.registry import platform_registry from integrations.chat_bridge.vendors.discord import discord_bot from integrations.telegram_bot.bot import telegram_bot await asyncio.sleep(0.5) # Register Discord in the platform registry platform_registry.register(discord_bot) if settings.telegram_token: try: await telegram_bot.start() logger.info("Telegram bot started") except Exception as exc: logger.warning("Failed to start Telegram bot: %s", exc) else: logger.debug("Telegram: no token configured, skipping") if settings.discord_token or discord_bot.load_token(): try: await discord_bot.start() logger.info("Discord bot started") except Exception as exc: logger.warning("Failed to start Discord bot: %s", exc) else: logger.debug("Discord: no token configured, skipping") # If Discord isn't connected yet, start a watcher that polls for the # token to appear in the environment or .env file. if discord_bot.state.name != "CONNECTED": asyncio.create_task(_discord_token_watcher()) async def _discord_token_watcher() -> None: """Poll for DISCORD_TOKEN appearing in env or .env and auto-start Discord bot.""" from integrations.chat_bridge.vendors.discord import discord_bot # Don't poll if discord.py isn't even installed try: import discord as _discord_check # noqa: F401 except ImportError: logger.debug("discord.py not installed — token watcher exiting") return while True: await asyncio.sleep(30) if discord_bot.state.name == "CONNECTED": return # Already running — stop watching # 1. Check settings (pydantic-settings reads env on instantiation; # hot-reload is handled by re-reading .env below) token = settings.discord_token # 2. Re-read .env file for hot-reload if not token: try: from dotenv import dotenv_values env_path = Path(settings.repo_root) / ".env" if env_path.exists(): vals = dotenv_values(env_path) token = vals.get("DISCORD_TOKEN", "") except ImportError: pass # python-dotenv not installed # 3. Check state file (written by /discord/setup) if not token: token = discord_bot.load_token() or "" if token: try: logger.info( "Discord watcher: token found, attempting start (state=%s)", discord_bot.state.name, ) success = await discord_bot.start(token=token) if success: logger.info("Discord bot auto-started (token detected)") return # Done — stop watching logger.warning( "Discord watcher: start() returned False (state=%s)", discord_bot.state.name, ) except Exception as exc: logger.warning("Discord auto-start failed: %s", exc) def _startup_init() -> None: """Validate config and enable event persistence.""" from config import validate_startup validate_startup() from infrastructure.events.bus import init_event_bus_persistence init_event_bus_persistence() from spark.engine import get_spark_engine if get_spark_engine().enabled: logger.info("Spark Intelligence active — event capture enabled") def _startup_background_tasks() -> list[asyncio.Task]: """Spawn all recurring background tasks (non-blocking).""" bg_tasks = [ asyncio.create_task(_briefing_scheduler()), asyncio.create_task(_thinking_scheduler()), asyncio.create_task(_loop_qa_scheduler()), asyncio.create_task(_presence_watcher()), asyncio.create_task(_start_chat_integrations_background()), ] try: from timmy.paperclip import start_paperclip_poller bg_tasks.append(asyncio.create_task(start_paperclip_poller())) logger.info("Paperclip poller started") except ImportError: logger.debug("Paperclip module not found, skipping poller") return bg_tasks def _try_prune(label: str, prune_fn, days: int) -> None: """Run a prune function, log results, swallow errors.""" try: pruned = prune_fn() if pruned: logger.info( "%s auto-prune: removed %d entries older than %d days", label, pruned, days, ) except Exception as exc: logger.debug("%s auto-prune skipped: %s", label, exc) def _check_vault_size() -> None: """Warn if the memory vault exceeds the configured size limit.""" try: vault_path = Path(settings.repo_root) / "memory" / "notes" if vault_path.exists(): total_bytes = sum(f.stat().st_size for f in vault_path.rglob("*") if f.is_file()) total_mb = total_bytes / (1024 * 1024) if total_mb > settings.memory_vault_max_mb: logger.warning( "Memory vault (%.1f MB) exceeds limit (%d MB) — consider archiving old notes", total_mb, settings.memory_vault_max_mb, ) except Exception as exc: logger.debug("Vault size check skipped: %s", exc) def _startup_pruning() -> None: """Auto-prune old memories, thoughts, and events on startup.""" if settings.memory_prune_days > 0: from timmy.memory_system import prune_memories _try_prune( "Memory", lambda: prune_memories( older_than_days=settings.memory_prune_days, keep_facts=settings.memory_prune_keep_facts, ), settings.memory_prune_days, ) if settings.thoughts_prune_days > 0: from timmy.thinking import thinking_engine _try_prune( "Thought", lambda: thinking_engine.prune_old_thoughts( keep_days=settings.thoughts_prune_days, keep_min=settings.thoughts_prune_keep_min, ), settings.thoughts_prune_days, ) if settings.events_prune_days > 0: from swarm.event_log import prune_old_events _try_prune( "Event", lambda: prune_old_events( keep_days=settings.events_prune_days, keep_min=settings.events_prune_keep_min, ), settings.events_prune_days, ) if settings.memory_vault_max_mb > 0: _check_vault_size() async def _shutdown_cleanup( bg_tasks: list[asyncio.Task], workshop_heartbeat, ) -> None: """Stop chat bots, MCP sessions, heartbeat, and cancel background tasks.""" from integrations.chat_bridge.vendors.discord import discord_bot from integrations.telegram_bot.bot import telegram_bot await discord_bot.stop() await telegram_bot.stop() try: from timmy.mcp_tools import close_mcp_sessions await close_mcp_sessions() except Exception as exc: logger.debug("MCP shutdown: %s", exc) await workshop_heartbeat.stop() for task in bg_tasks: task.cancel() try: await task except asyncio.CancelledError: pass @asynccontextmanager async def lifespan(app: FastAPI): """Application lifespan manager with non-blocking startup.""" _startup_init() bg_tasks = _startup_background_tasks() _startup_pruning() # Start Workshop presence heartbeat with WS relay from dashboard.routes.world import broadcast_world_state from timmy.workshop_state import WorkshopHeartbeat workshop_heartbeat = WorkshopHeartbeat(on_change=broadcast_world_state) await workshop_heartbeat.start() # Register session logger with error capture try: from infrastructure.error_capture import register_error_recorder from timmy.session_logger import get_session_logger register_error_recorder(get_session_logger().record_error) except Exception: logger.debug("Failed to register error recorder") logger.info("✓ Dashboard ready for requests") yield await _shutdown_cleanup(bg_tasks, workshop_heartbeat) app = FastAPI( title="Mission Control", version="1.0.0", lifespan=lifespan, docs_url="/docs", openapi_url="/openapi.json", ) def _get_cors_origins() -> list[str]: """Get CORS origins from settings, rejecting wildcards in production. Adds matrix_frontend_url when configured. Always allows Tailscale IPs (100.x.x.x range) for development convenience. """ origins = list(settings.cors_origins) # Strip wildcards in production (security) if "*" in origins and not settings.debug: logger.warning( "Wildcard '*' in CORS_ORIGINS stripped in production — " "set explicit origins via CORS_ORIGINS env var" ) origins = [o for o in origins if o != "*"] # Add Matrix frontend URL if configured if settings.matrix_frontend_url: url = settings.matrix_frontend_url.strip() if url and url not in origins: origins.append(url) logger.debug("Added Matrix frontend to CORS: %s", url) return origins # Pattern to match Tailscale IPs (100.x.x.x) for CORS origin regex _TAILSCALE_IP_PATTERN = re.compile(r"^https?://100\.\d{1,3}\.\d{1,3}\.\d{1,3}(?::\d+)?$") def _is_tailscale_origin(origin: str) -> bool: """Check if origin is a Tailscale IP (100.x.x.x range).""" return bool(_TAILSCALE_IP_PATTERN.match(origin)) # Add dedicated middleware in correct order # 1. Logging (outermost to capture everything) app.add_middleware(RequestLoggingMiddleware, skip_paths=["/health"]) # 2. Rate Limiting (before security to prevent abuse early) app.add_middleware( RateLimitMiddleware, path_prefixes=["/api/matrix/"], requests_per_minute=30, ) # 3. Security Headers app.add_middleware(SecurityHeadersMiddleware, production=not settings.debug) # 4. CSRF Protection app.add_middleware(CSRFMiddleware) # 4. Standard FastAPI middleware # In development, allow all hosts (Tailscale IPs, MagicDNS, etc.) _trusted = settings.trusted_hosts if settings.timmy_env == "production" else ["*"] app.add_middleware( TrustedHostMiddleware, allowed_hosts=_trusted, ) app.add_middleware( CORSMiddleware, allow_origins=_get_cors_origins(), allow_origin_regex=r"https?://100\.\d{1,3}\.\d{1,3}\.\d{1,3}(:\d+)?", allow_credentials=True, allow_methods=["GET", "POST", "PUT", "DELETE", "OPTIONS"], allow_headers=["Content-Type", "Authorization"], ) # Mount static files static_dir = PROJECT_ROOT / "static" if static_dir.exists(): app.mount("/static", StaticFiles(directory=str(static_dir)), name="static") # Shared templates instance from dashboard.templating import templates # noqa: E402 # Include routers app.include_router(health_router) app.include_router(agents_router) app.include_router(voice_router) app.include_router(mobile_router) app.include_router(briefing_router) app.include_router(telegram_router) app.include_router(tools_router) app.include_router(spark_router) app.include_router(discord_router) app.include_router(memory_router) app.include_router(grok_router) app.include_router(models_router) app.include_router(models_api_router) app.include_router(chat_api_router) app.include_router(chat_api_v1_router) app.include_router(thinking_router) app.include_router(calm_router) app.include_router(tasks_router) app.include_router(work_orders_router) app.include_router(loop_qa_router) app.include_router(system_router) app.include_router(experiments_router) app.include_router(db_explorer_router) app.include_router(world_router) app.include_router(matrix_router) app.include_router(tower_router) app.include_router(daily_run_router) app.include_router(quests_router) app.include_router(scorecards_router) app.include_router(sovereignty_metrics_router) @app.websocket("/ws") async def ws_redirect(websocket: WebSocket): """Catch stale /ws connections and close cleanly. websockets 16.0 dropped the legacy ``transfer_data_task`` attribute, so calling ``websocket.close()`` after accept triggers an AttributeError. Use the raw ASGI send instead. """ await websocket.accept() try: await websocket.close(code=1008, reason="Deprecated endpoint") except AttributeError: # websockets >= 16.0 — close via raw ASGI message await websocket.send({"type": "websocket.close", "code": 1008}) @app.websocket("/swarm/live") async def swarm_live(websocket: WebSocket): """Swarm live event stream via WebSocket.""" from infrastructure.ws_manager.handler import ws_manager as ws_mgr await ws_mgr.connect(websocket) try: while True: # Keep connection alive; events are pushed via ws_mgr.broadcast() await websocket.receive_text() except Exception as exc: logger.debug("WebSocket disconnect error: %s", exc) ws_mgr.disconnect(websocket) @app.get("/swarm/agents/sidebar", response_class=HTMLResponse) async def swarm_agents_sidebar(): """HTMX partial: list active swarm agents for the dashboard sidebar.""" try: from config import settings agents_yaml = settings.agents_config agents = agents_yaml.get("agents", {}) lines = [] for name, cfg in agents.items(): model = cfg.get("model", "default") lines.append( f'
' f'{name}' f'{model}' f"
" ) return "\n".join(lines) if lines else '
No agents configured
' except Exception as exc: logger.debug("Agents sidebar error: %s", exc) return '
Agents unavailable
' @app.get("/", response_class=HTMLResponse) async def root(request: Request): """Serve the main dashboard page.""" return templates.TemplateResponse(request, "index.html", {}) @app.get("/shortcuts/setup") async def shortcuts_setup(): """Siri Shortcuts setup guide.""" from integrations.shortcuts.siri import get_setup_guide return get_setup_guide()