- scripts/provision-timmy-vps.sh: Full automated provisioning - configs/llama-server.service: Inference systemd unit - configs/timmy-agent.service: Agent harness systemd unit - docs/VPS_SETUP.md: Setup and troubleshooting guide Installs llama.cpp, Hermes-3 model, Python venv, firewall rules. Configures localhost-only inference on port 8081.
23 lines
412 B
Desktop File
23 lines
412 B
Desktop File
[Unit]
|
|
Description=llama.cpp inference server for Timmy
|
|
After=network.target
|
|
|
|
[Service]
|
|
Type=simple
|
|
User=root
|
|
WorkingDirectory=/root/timmy
|
|
ExecStart=/root/timmy/llama-server \
|
|
-m /root/timmy/models/hermes-3-8b.Q4_K_M.gguf \
|
|
--host 127.0.0.1 \
|
|
--port 8081 \
|
|
-c 8192 \
|
|
-np 1 \
|
|
--jinja \
|
|
-ngl 0
|
|
Restart=always
|
|
RestartSec=10
|
|
Environment="HOME=/root"
|
|
|
|
[Install]
|
|
WantedBy=multi-user.target
|