Files
timmy-config/allegro/docs/VPS_SETUP.md
2026-03-31 20:02:01 +00:00

2.6 KiB

Timmy VPS Setup Guide

Complete guide for provisioning a sovereign Timmy wizard VPS.

Prerequisites

  • Fresh Ubuntu 22.04 or 24.04 VPS
  • Root SSH access
  • At least 4GB RAM, 20GB disk
  • Internet connection

Quick Start

# Download and run provisioning script
curl -sL https://raw.githubusercontent.com/Timmy_Foundation/timmy-home/main/scripts/provision-timmy-vps.sh | bash

What Gets Installed

Component Purpose Port
llama.cpp Local inference 8081 (localhost only)
Python venv Agent environment -
timmy-home Agent scripts -
Syncthing File sync 22000
UFW Firewall -

Directory Structure

~/timmy/
├── models/          # AI model weights
├── soul/            # Conscience files (SOUL.md)
├── scripts/         # Operational scripts
├── logs/            # Agent logs
├── shared/          # Syncthing shared folder
├── configs/         # Configuration files
└── timmy-home/      # Repository clone

Services

llama-server

Local inference server (CPU-only with OpenBLAS)

systemctl status llama-server
systemctl restart llama-server
journalctl -u llama-server -f

timmy-agent

Agent harness that calls local inference

systemctl status timmy-agent
systemctl restart timmy-agent

syncthing

File synchronization between VPS nodes

systemctl status syncthing@root

Testing

Check Inference

curl http://127.0.0.1:8081/health

Test Completion

curl -X POST http://127.0.0.1:8081/completion \
  -H "Content-Type: application/json" \
  -d '{"prompt": "Hello, I am", "max_tokens": 10}'

System Status

~/timmy/scripts/status.sh

Security

  • Port 8081 (inference): localhost only, never exposed
  • Port 22000 (syncthing): open for P2P sync
  • Port 22 (SSH): standard access
  • UFW: All other ports blocked by default

Troubleshooting

Issue Solution
llama-server won't start Check model exists: ls ~/timmy/models/
Out of memory Use smaller GGUF (Q4_K_S instead of Q4_K_M)
Syncthing not syncing Check firewall: ufw status
Slow inference Ensure OpenBLAS is working: ldd ~/timmy/llama-server | grep blas

Manual Model Download

If automatic download fails:

cd ~/timmy/models
wget https://huggingface.co/TheBloke/Hermes-3-Llama-3.1-8B-GGUF/resolve/main/hermes-3-llama-3.1-8b.Q4_K_M.gguf
systemctl restart llama-server

Uninstall

systemctl stop llama-server timmy-agent syncthing@root
systemctl disable llama-server timmy-agent syncthing@root
rm -rf ~/timmy