Files
timmy-home/configs/llama-server.service
Allegro 266d6ec008 [#75] Add VPS provisioning script for sovereign Timmy deployment
- scripts/provision-timmy-vps.sh: Full automated provisioning
- configs/llama-server.service: Inference systemd unit
- configs/timmy-agent.service: Agent harness systemd unit
- docs/VPS_SETUP.md: Setup and troubleshooting guide

Installs llama.cpp, Hermes-3 model, Python venv, firewall rules.
Configures localhost-only inference on port 8081.
2026-03-30 15:22:34 +00:00

23 lines
412 B
Desktop File

[Unit]
Description=llama.cpp inference server for Timmy
After=network.target
[Service]
Type=simple
User=root
WorkingDirectory=/root/timmy
ExecStart=/root/timmy/llama-server \
-m /root/timmy/models/hermes-3-8b.Q4_K_M.gguf \
--host 127.0.0.1 \
--port 8081 \
-c 8192 \
-np 1 \
--jinja \
-ngl 0
Restart=always
RestartSec=10
Environment="HOME=/root"
[Install]
WantedBy=multi-user.target