2026-02-19 19:05:01 +00:00
# Timmy Time — Mission Control
2026-02-19 19:11:42 +00:00
A local-first dashboard for your sovereign AI agents. Talk to Timmy, watch his status, verify Ollama is running — all from a browser, no cloud required.
2026-02-19 19:05:01 +00:00
2026-02-19 19:11:42 +00:00
---
2026-02-19 19:05:01 +00:00
2026-02-19 19:11:42 +00:00
## Prerequisites
2026-02-19 19:05:01 +00:00
2026-02-19 19:11:42 +00:00
You need three things on your Mac before anything else:
2026-02-19 19:05:01 +00:00
2026-02-19 19:11:42 +00:00
**Python 3.11+**
2026-02-19 19:05:01 +00:00
```bash
2026-02-19 19:11:42 +00:00
python3 --version # should be 3.11 or higher
```
If not: `brew install python@3.11`
**Ollama** (runs the local LLM)
```bash
brew install ollama
```
Or download from https://ollama.com
**Git** — already on every Mac.
---
## Quickstart (copy-paste friendly)
### 1. Clone the branch
```bash
git clone -b claude/run-tests-IYl0F https://github.com/Alexspayne/Timmy-time-dashboard.git
cd Timmy-time-dashboard
```
### 2. Create a virtual environment and install
```bash
python3 -m venv .venv
source .venv/bin/activate
2026-02-19 19:05:01 +00:00
pip install -e ".[dev]"
2026-02-19 19:11:42 +00:00
```
### 3. Pull the model (one-time, ~2 GB download)
Open a **new terminal tab ** and run:
```bash
ollama serve
```
Back in your first tab:
```bash
ollama pull llama3.2
```
2026-02-19 19:05:01 +00:00
2026-02-19 19:11:42 +00:00
### 4. Start the dashboard
2026-02-19 19:05:01 +00:00
2026-02-19 19:11:42 +00:00
```bash
2026-02-19 19:05:01 +00:00
uvicorn dashboard.app:app --reload
2026-02-19 19:11:42 +00:00
```
Open your browser to **http://localhost:8000 **
---
## What you'll see
The dashboard has two panels on the left and a chat window on the right:
- **AGENTS** — Timmy's metadata (model, type, version)
- **SYSTEM HEALTH** — live Ollama status, auto-refreshes every 30 seconds
- **TIMMY INTERFACE** — type a message, hit SEND, get a response from the local LLM
2026-02-19 19:05:01 +00:00
2026-02-19 19:11:42 +00:00
If Ollama isn't running when you send a message, the chat will show a "Timmy is offline" error instead of crashing.
---
## Run the tests
No Ollama needed — all external calls are mocked.
```bash
2026-02-19 19:05:01 +00:00
pytest
```
2026-02-19 19:11:42 +00:00
Expected output:
```
27 passed in 0.67s
```
---
## Optional: CLI
With your venv active:
2026-02-19 19:05:01 +00:00
```bash
timmy chat "What is sovereignty?"
timmy think "Bitcoin and self-custody"
timmy status
```
2026-02-19 19:11:42 +00:00
---
## Project layout
2026-02-19 19:05:01 +00:00
```
src/
2026-02-19 19:11:42 +00:00
timmy/ # Timmy agent — wraps Agno (soul = prompt, body = Agno)
dashboard/ # FastAPI app + routes + Jinja2 templates
static/ # CSS (dark mission-control theme)
tests/ # 27 pytest tests
pyproject.toml # dependencies and build config
2026-02-19 19:05:01 +00:00
```
2026-02-19 19:11:42 +00:00
---
## Troubleshooting
**`ollama: command not found` ** — Ollama isn't installed or isn't on your PATH. Install via Homebrew or the .dmg from ollama.com.
**`connection refused` in the chat** — Ollama isn't running. Open a terminal and run `ollama serve` , then try again.
**`ModuleNotFoundError: No module named 'dashboard'` ** — You're not in the venv or forgot `pip install -e .` . Run `source .venv/bin/activate` then `pip install -e ".[dev]"` .
**Health panel shows DOWN** — Ollama isn't running. The chat still works for testing but will return the offline error message.
---
2026-02-19 19:05:01 +00:00
## Roadmap
| Version | Name | Milestone |
|---------|------------|--------------------------------------------|
| 1.0.0 | Genesis | Agno + Ollama + SQLite + Dashboard |
| 2.0.0 | Exodus | MCP tools + multi-agent |
| 3.0.0 | Revelation | Bitcoin Lightning treasury + single `.app` |