11 KiB
GENOME.md — timmy-config
Auto-generated facts were derived from the local ~/.timmy/timmy-config checkout on 2026-04-14 and then reviewed manually for architecture, API surface, and operational meaning.
Project Overview
timmy-config is the sovereign sidecar for Timmy.
It is not the model runtime and it is not the work archive. It is the configuration and orchestration layer that tells Hermes who Timmy is, how he is routed, what scripts are available, what memories and skins are installed, and how the fleet is operated.
The repo exists because the Timmy Foundation made a hard architectural split:
hermes-agentis the enginetimmy-configis the driver's seattimmy-homeis the lived workspace and artifact archive
This repo is therefore a mixed codebase with three major responsibilities:
- Sidecar deployment into
~/.hermes/ - Operational automation for the fleet
- Thin training/orchestration support without becoming a monolith again
Observed quick facts from the current local checkout:
- Source files: 178
- Test files: 36
- Config files: 24
- Total lines: 35,624
- Last commit on analyzed source:
7630806f(sync: align repo with live system config) - Total commits: 393
Architecture Diagram
flowchart TD
A[timmy-config repo] --> B[deploy.sh]
B --> C[~/.hermes/config.yaml]
B --> D[~/.hermes/bin/*]
B --> E[~/.hermes/playbooks/*]
B --> F[~/.hermes/cron/*]
B --> G[~/.hermes/memories/*]
B --> H[~/.timmy/SOUL.md]
C --> I[Hermes gateway/runtime]
D --> J[Operational scripts]
E --> K[Playbook-driven agent behavior]
F --> L[Cron-scheduled automation]
J --> M[Gitea forge]
J --> N[Telegram / platform gateways]
J --> O[tmux wizard fleet]
J --> P[metrics + local files]
Q[orchestration.py + tasks.py] --> R[Huey / SQLite queue]
R --> J
Q --> M
Q --> P
S[gitea_client.py] --> M
T[training/*] --> U[curated datasets + evaluation recipes]
U --> P
V[timmy-home] <-->|artifacts / logs / training outputs| J
V <-->|continuity / notes / metrics| Q
Entry Points and Data Flow
Primary entry points
-
deploy.sh- canonical sidecar deployment entry point
- copies config, scripts, playbooks, cron definitions, skins, and memories into
~/.hermes/ - writes
SOUL.mdinto~/.timmy/
-
orchestration.py- defines the Huey SQLite queue
- the root scheduler primitive for queued work
-
tasks.py- the real orchestration surface
- contains scheduled jobs, local Hermes invocations, archive/training helpers, JSONL helpers, continuity flushing, and repo lists
-
gitea_client.py- typed stdlib-only interface to Gitea
- shared API layer replacing scattered raw curl logic
-
bin/*- operational executables for health checks, deadman switch, dispatch, watchdogs, scans, and status panels
- this is the hands-on operator layer
Data flow summary
- Configuration starts in repo files like
config.yaml,fallback-portfolios.yaml,channel_directory.json,memories/,skins/,playbooks/, andcron/ deploy.shoverlays these into the Hermes runtime directory- Hermes runtime reads the deployed config and scripts
tasks.pyandbin/*then interact with:- Gitea (
gitea_client.py, issue/PR automation) - local Hermes sessions (
run_hermes_localpaths intasks.py) - local files in
~/.timmy/and~/.hermes/ - metrics JSONL outputs
- Telegram / gateway surfaces
- Gitea (
This is a classic sidecar pattern: the repo does not own the engine, but it owns almost every operational decision around the engine.
Key Abstractions
1. Sidecar overlay
Core idea: never fork Hermes if a sidecar can express the behavior. This abstraction appears in:
deploy.sh- repo layout under
bin/,cron/,playbooks/,memories/,skins/ - the explicit README boundary between
timmy-configandtimmy-home
2. Typed forge client
gitea_client.py turns the forge into a stable internal API surface.
Important abstractions:
IssuePullRequestCommentLabelUserGiteaClient
This is important because many other scripts can depend on one client instead of shelling out to brittle curl commands.
3. Queue-backed orchestration
orchestration.py + tasks.py define the move from ad-hoc shell automation to queued work.
The central abstraction is not just “a cron job” but “a schedulable task with local model execution, continuity, and metrics.”
4. Continuity as files, not vibes
tasks.py contains explicit file-backed continuity helpers:
flush_continuity(...)- JSON readers/writers
- JSONL append/load helpers
- archive checkpoint/state files
This abstraction matters because Timmy continuity survives compaction or restart by being written to disk.
5. Training as thin recipes
The training directory is intentionally framed as transitional. It exposes recipes and helper scripts, but the README is explicit that lived data belongs elsewhere. This is an important abstraction boundary:
- configs + generators here
- real activity artifacts in
timmy-home
API Surface
Shell / operator API
Important user-facing commands implied by the repo:
./deploy.sh- Huey consumer startup via
huey_consumer.py tasks.huey -w 2 -k thread - scripts in
bin/such as:deadman-switch.shfleet-status.shmodel-health-check.shstart-loops.shagent-dispatch.sh
Python API surface
Most reusable programmatic interfaces:
gitea_client.GiteaClient
Key methods include:
list_org_reposlist_issuesget_issuecreate_issueupdate_issueclose_issueassign_issueadd_labelslist_commentscreate_commentlist_pullsget_pullcreate_pullmerge_pullupdate_pull_branchclose_pull
tasks.py helpers
Notable reusable surfaces:
run_hermes_local(...)hermes_local(...)run_reflex_task(...)run_archive_hermes(...)flush_continuity(...)- JSON/JSONL primitives:
read_jsonwrite_jsonload_jsonlwrite_jsonlappend_jsonlcount_jsonl_rows
This API surface is broad enough that timmy-config is functionally an operations SDK as much as a config repo.
Test Coverage Gaps
The codebase-genome pipeline estimated:
- source modules: 86
- test modules: 36
- estimated coverage: 15%
- untested modules: 73
That estimate is crude, but directionally useful.
High-value gaps
The most important untested or under-tested areas are not random scripts. They are the system boundary scripts and operator-critical surfaces:
-
deploy.sh- highest leverage file in the repo
- if deploy breaks, the sidecar stops being the source of truth in practice
-
orchestration.pyand large portions oftasks.py- especially scheduling, local-Hermes execution wrappers, checkpoint/state flows, and failure modes
-
gitea_client.py- some behavior may be covered indirectly, but the client is important enough to deserve deeper contract tests for pagination, merge failures, 405 handling, and retry behavior
-
bin/*operational scripts- deadman, watchdog, model health, and dispatch paths are safety critical
-
Ansible deployment surface
ansible/playbooks/*andansible/roles/*are configuration-heavy but still represent production behavior
Practical takeaway
If the next wave of testing work is prioritized, the order should be:
- deploy + runtime overlay correctness
- Gitea client contracts
- queue/task execution behavior in
tasks.py - deadman / watchdog / fleet-health scripts
- training and archive helper edge cases
Security Considerations
The pipeline already flags several categories, and manual review supports them.
1. Subprocess-heavy code
This repo uses shell and subprocess execution widely.
That is expected for an ops repo, but it increases command-injection risk.
Files like tasks.py, deployment helpers, and many bin/* scripts need careful boundary checking around interpolated inputs.
2. Secret adjacency
This repo references tokens, auth files, routing configs, and platform integrations. Even when secrets are not hardcoded, the repo lives close to sensitive paths. That means review discipline matters:
- no accidental dumps of live tokens
- no committing generated auth artifacts
- no relaxed assumptions about local paths being safe to expose
3. HTTP and webhook surfaces
Multiple scripts make outbound HTTP calls or serve automation endpoints. This means input validation, response validation, and timeout/error handling matter more than in a static config repo.
4. SQLite and file-backed state
SQLite/Huey plus JSON/JSONL file state are simple and sovereign, but they create corruption and stale-state risks if writes are partial or multiple writers race. The repo already encodes a strong preference for explicit checkpoints and continuity flushes, which is the right direction.
5. Sidecar privilege
Because deploy.sh writes directly into ~/.hermes/, this repo effectively has configuration authority over the live runtime.
That is powerful and dangerous. A bad deploy can break routing, scripts, or identity. In security terms, this repo is a control plane.
Performance Characteristics
timmy-config is not performance-sensitive in the same way a game loop or serving stack is. Its performance profile is operational:
- Many small scripts, low-latency startup expectations
- File-backed state and JSONL append patterns optimized for simplicity over throughput
- Huey + SQLite chosen for low operational overhead, not horizontal scale
- Deployment is copy-based and cheap; correctness matters more than speed
- The repo is large in surface area but shallow in runtime residency — most code runs only when called
Practical performance characteristics:
- deploy path should remain fast because it is mostly file copies
- Gitea automation cost is network-bound, not CPU-bound
- training helpers are the heaviest local operations but intentionally limited in scope
- orchestration latency is dominated by external model/API calls rather than Python compute
The biggest performance risk is not CPU. It is operational drift: too many scripts, overlapping logic, or old paths that still appear to exist. That increases human debugging cost even when machine runtime is fine.
Final Assessment
timmy-config is best understood as a sovereign control plane.
It is part config repo, part operations toolkit, part lightweight SDK, and part deployment overlay.
Its strongest architectural idea is the boundary:
- keep Hermes upstream
- keep Timmy sovereign through the sidecar
- keep lived work in timmy-home
Its biggest risk is sprawl. A repo with 178 source files and 393 commits can easily become the place where every operational idea goes to live forever. The right long-term direction is not to make it smaller by deleting its authority, but to keep sharpening the boundaries inside it:
- sidecar deploy surface
- queue/orchestration surface
- Gitea client surface
- training recipe surface
- operator script surface
That keeps the genome legible. Without that, the repo becomes powerful but opaque — and a control plane cannot afford opacity.