Step-by-step clone → venv → install → Ollama → run sequence, what to expect in the browser, troubleshooting section for common Mac failure modes. https://claude.ai/code/session_01M4L3R98N5fgXFZRvV8X9b6
Timmy Time — Mission Control
A local-first dashboard for your sovereign AI agents. Talk to Timmy, watch his status, verify Ollama is running — all from a browser, no cloud required.
Prerequisites
You need three things on your Mac before anything else:
Python 3.11+
python3 --version # should be 3.11 or higher
If not: brew install python@3.11
Ollama (runs the local LLM)
brew install ollama
Or download from https://ollama.com
Git — already on every Mac.
Quickstart (copy-paste friendly)
1. Clone the branch
git clone -b claude/run-tests-IYl0F https://github.com/Alexspayne/Timmy-time-dashboard.git
cd Timmy-time-dashboard
2. Create a virtual environment and install
python3 -m venv .venv
source .venv/bin/activate
pip install -e ".[dev]"
3. Pull the model (one-time, ~2 GB download)
Open a new terminal tab and run:
ollama serve
Back in your first tab:
ollama pull llama3.2
4. Start the dashboard
uvicorn dashboard.app:app --reload
Open your browser to http://localhost:8000
What you'll see
The dashboard has two panels on the left and a chat window on the right:
- AGENTS — Timmy's metadata (model, type, version)
- SYSTEM HEALTH — live Ollama status, auto-refreshes every 30 seconds
- TIMMY INTERFACE — type a message, hit SEND, get a response from the local LLM
If Ollama isn't running when you send a message, the chat will show a "Timmy is offline" error instead of crashing.
Run the tests
No Ollama needed — all external calls are mocked.
pytest
Expected output:
27 passed in 0.67s
Optional: CLI
With your venv active:
timmy chat "What is sovereignty?"
timmy think "Bitcoin and self-custody"
timmy status
Project layout
src/
timmy/ # Timmy agent — wraps Agno (soul = prompt, body = Agno)
dashboard/ # FastAPI app + routes + Jinja2 templates
static/ # CSS (dark mission-control theme)
tests/ # 27 pytest tests
pyproject.toml # dependencies and build config
Troubleshooting
ollama: command not found — Ollama isn't installed or isn't on your PATH. Install via Homebrew or the .dmg from ollama.com.
connection refused in the chat — Ollama isn't running. Open a terminal and run ollama serve, then try again.
ModuleNotFoundError: No module named 'dashboard' — You're not in the venv or forgot pip install -e .. Run source .venv/bin/activate then pip install -e ".[dev]".
Health panel shows DOWN — Ollama isn't running. The chat still works for testing but will return the offline error message.
Roadmap
| Version | Name | Milestone |
|---|---|---|
| 1.0.0 | Genesis | Agno + Ollama + SQLite + Dashboard |
| 2.0.0 | Exodus | MCP tools + multi-agent |
| 3.0.0 | Revelation | Bitcoin Lightning treasury + single .app |