Compare commits

..

2 Commits

Author SHA1 Message Date
Alexander Whitestone
06a6b53a37 feat: add tower world-changing items for #514
Some checks failed
Agent PR Gate / gate (pull_request) Failing after 58s
Self-Healing Smoke / self-healing-smoke (pull_request) Failing after 26s
Agent PR Gate / report (pull_request) Successful in 23s
Smoke Test / smoke (pull_request) Failing after 27s
2026-04-22 02:54:56 -04:00
Alexander Whitestone
5ce4fa6cb8 test: define tower world items for #514 2026-04-22 02:43:29 -04:00
5 changed files with 219 additions and 183 deletions

View File

@@ -12,6 +12,29 @@ WORLD_DIR = Path('/Users/apayne/.timmy/evennia/timmy_world')
STATE_FILE = WORLD_DIR / 'game_state.json'
TIMMY_LOG = WORLD_DIR / 'timmy_log.md'
WORLD_ITEMS = {
"foraged key": {"effect": "unlock_tower_cache", "quest_item": True, "consumable": False, "effect_text": "A hidden cache clicks open in the Tower wall."},
"seed packet": {"effect": "grow_garden", "quest_item": False, "consumable": True, "effect_text": "Fresh growth pushes through the Garden soil."},
"notebook": {"effect": "write_notebook_rule", "quest_item": False, "consumable": False, "effect_text": "A new rule joins the whiteboard in the Tower."},
"cloth": {"effect": "patch_bridge", "quest_item": False, "consumable": True, "effect_text": "The Bridge railing is wrapped tight against the weather."},
"oil can": {"effect": "stoke_forge", "quest_item": False, "consumable": True, "effect_text": "The Forge fire answers with a hotter glow."},
"lantern": {"effect": "light_bridge", "quest_item": False, "consumable": False, "effect_text": "A steady lantern glow cuts through the dark over the Bridge."},
"rope spool": {"effect": "secure_bridge", "quest_item": False, "consumable": True, "effect_text": "The Bridge is lashed tight and feels safer underfoot."},
"chalk": {"effect": "mark_threshold", "quest_item": False, "consumable": True, "effect_text": "A chalk mark at the Threshold points wanderers home."},
"weather vane": {"effect": "read_weather", "quest_item": False, "consumable": False, "effect_text": "The weather vane settles and the coming storm makes sense."},
"sunstone": {"effect": "restore_tower_power", "quest_item": False, "consumable": False, "effect_text": "Warm light races through the Tower circuits."},
"iron nails": {"effect": "reinforce_bridge", "quest_item": False, "consumable": True, "effect_text": "The Bridge planks are pinned down against the flood."},
"river stone": {"effect": "water_garden", "quest_item": False, "consumable": True, "effect_text": "Moisture returns to the Garden beds."},
}
ROOM_DISCOVERABLES = {
"Threshold": ["chalk", "sunstone"],
"Tower": ["notebook", "lantern"],
"Forge": ["oil can", "iron nails"],
"Garden": ["seed packet", "foraged key"],
"Bridge": ["cloth", "rope spool", "weather vane", "river stone"],
}
# ============================================================
# NARRATIVE ARC — 4 phases that transform the world
# ============================================================
@@ -143,6 +166,8 @@ class World:
"visitors": [],
},
}
for room_name, items in ROOM_DISCOVERABLES.items():
self.rooms[room_name]["discoverables"] = list(items)
# Characters (not NPCs — they have lives)
self.characters = {
@@ -258,6 +283,14 @@ class World:
"items_crafted": 0,
"conflicts_resolved": 0,
"nights_survived": 0,
"bridge_patched": False,
"bridge_secured": False,
"bridge_reinforced": False,
"bridge_lantern_lit": False,
"tower_cache_unlocked": False,
"threshold_marked": False,
"weather_readable": False,
"sunstone_socketed": False,
}
def tick_time(self):
@@ -376,6 +409,14 @@ class World:
desc += " Rain mists on the dark water below."
if len(self.rooms["Bridge"]["carvings"]) > 1:
desc += f" There are {len(self.rooms['Bridge']['carvings'])} carvings now."
if self.state.get("bridge_patched"):
desc += " Cloth bindings keep the railing from splintering."
if self.state.get("bridge_secured"):
desc += " Rope lines keep the span steady against the flood."
if self.state.get("bridge_reinforced"):
desc += " Fresh iron nails hold the planks tight."
if self.state.get("bridge_lantern_lit"):
desc += " A lantern glows warm over the water."
elif room_name == "Tower":
power = self.state.get("tower_power_low", False)
@@ -384,6 +425,19 @@ class World:
if self.rooms["Tower"]["messages"]:
desc += f" The whiteboard holds {len(self.rooms['Tower']['messages'])} rules."
if self.state.get("tower_cache_unlocked"):
desc += " A hidden cache stands open beneath the whiteboard."
if room_name == "Threshold" and self.state.get("threshold_marked"):
desc += " A chalk arrow points late arrivals toward shelter."
if room_name == "Garden" and self.state.get("weather_readable"):
desc += " The beds are arranged to catch whatever weather comes next."
if room_name == "Tower" and self.state.get("sunstone_socketed"):
desc += " A sunstone keeps the room lit with a stubborn amber glow."
discoverables = room.get("discoverables", [])
if discoverables:
desc += f" Discoverable items: {', '.join(discoverables)}."
# Who's here
here = [n for n, c in self.characters.items() if c["room"] == room_name and n != char_name]
@@ -485,6 +539,10 @@ class ActionSystem:
"cost": 1,
"description": "Take an item from the room",
},
"use": {
"cost": 1,
"description": "Use an item from your inventory to change the world",
},
"examine": {
"cost": 0,
"description": "Examine something in detail",
@@ -530,6 +588,13 @@ class ActionSystem:
available.append("rest")
available.append("examine")
discoverables = world.rooms[room].get("discoverables", [])
for item in discoverables:
available.append(f"take:{item}")
for item in char["inventory"]:
available.append(f"use:{item}")
if char["inventory"]:
available.append("give:item")
@@ -1072,6 +1137,76 @@ class GameEngine:
f.write(f"\n*Began: {datetime.now().strftime('%Y-%m-%d %H:%M')}*\n\n")
f.write("---\n\n")
f.write(message + "\n")
def _take_item(self, item_name, scene):
room_name = self.world.characters["Timmy"]["room"]
discoverables = self.world.rooms[room_name].get("discoverables", [])
if item_name not in discoverables:
scene["log"].append(f"There is no {item_name} here.")
return
discoverables.remove(item_name)
self.world.characters["Timmy"]["inventory"].append(item_name)
scene["log"].append(f"You take the {item_name}.")
if WORLD_ITEMS.get(item_name, {}).get("quest_item"):
scene["world_events"].append(f"The {item_name} feels important. It might open a quest route.")
def _use_item(self, item_name, scene):
inventory = self.world.characters["Timmy"]["inventory"]
if item_name not in inventory:
scene["log"].append(f"You are not carrying {item_name}.")
return
item = WORLD_ITEMS.get(item_name)
if not item:
scene["log"].append(f"The {item_name} doesn't seem to do anything.")
return
effect = item["effect"]
effect_text = item["effect_text"]
if effect == "grow_garden":
self.world.rooms["Garden"]["growth"] = min(5, self.world.rooms["Garden"]["growth"] + 2)
self.world.state["garden_drought"] = False
elif effect == "unlock_tower_cache":
self.world.state["tower_cache_unlocked"] = True
cache_rule = "Rule: Keys open more than doors when the world trusts you."
if cache_rule not in self.world.rooms["Tower"]["messages"]:
self.world.rooms["Tower"]["messages"].append(cache_rule)
elif effect == "write_notebook_rule":
note_rule = f"Rule #{self.world.tick}: A notebook can turn memory into structure."
self.world.rooms["Tower"]["messages"].append(note_rule)
elif effect == "patch_bridge":
self.world.state["bridge_patched"] = True
self.world.state["bridge_flooding"] = False
self.world.rooms["Bridge"]["weather"] = None
self.world.rooms["Bridge"]["rain_ticks"] = 0
elif effect == "stoke_forge":
self.world.rooms["Forge"]["fire"] = "glowing"
self.world.state["forge_fire_dying"] = False
elif effect == "light_bridge":
self.world.state["bridge_lantern_lit"] = True
elif effect == "secure_bridge":
self.world.state["bridge_secured"] = True
self.world.state["bridge_flooding"] = False
self.world.rooms["Bridge"]["weather"] = None
self.world.rooms["Bridge"]["rain_ticks"] = 0
elif effect == "mark_threshold":
self.world.state["threshold_marked"] = True
elif effect == "read_weather":
self.world.state["weather_readable"] = True
self.world.state["garden_drought"] = False
elif effect == "restore_tower_power":
self.world.state["tower_power_low"] = False
self.world.state["sunstone_socketed"] = True
elif effect == "reinforce_bridge":
self.world.state["bridge_reinforced"] = True
self.world.state["bridge_flooding"] = False
elif effect == "water_garden":
self.world.state["garden_drought"] = False
self.world.rooms["Garden"]["growth"] = min(5, self.world.rooms["Garden"]["growth"] + 1)
scene["log"].append(f"You use the {item_name}. {effect_text}")
scene["world_events"].append(effect_text)
if item.get("consumable"):
inventory.remove(item_name)
def run_tick(self, timmy_action="look"):
"""Run one tick. Return the scene and available choices."""
@@ -1100,7 +1235,7 @@ class GameEngine:
action_costs = {
"move": 2, "tend_fire": 3, "write_rule": 2, "carve": 2,
"plant": 2, "study": 2, "forge": 3, "help": 2, "speak": 1,
"listen": 0, "rest": -2, "examine": 0, "give": 0, "take": 1,
"listen": 0, "rest": -2, "examine": 0, "give": 0, "take": 1, "use": 1,
}
# Extract action name
@@ -1360,9 +1495,17 @@ class GameEngine:
elif timmy_action == "examine":
room = self.world.characters["Timmy"]["room"]
room_data = self.world.rooms[room]
items = room_data.get("items", [])
items = room_data.get("items", []) + room_data.get("discoverables", [])
scene["log"].append(f"You examine The {room}. You see: {', '.join(items) if items else 'nothing special'}")
elif timmy_action.startswith("take:"):
item_name = timmy_action.split(":", 1)[1]
self._take_item(item_name, scene)
elif timmy_action.startswith("use:"):
item_name = timmy_action.split(":", 1)[1]
self._use_item(item_name, scene)
elif timmy_action.startswith("help:"):
# Help increases trust
target_name = timmy_action.split(":")[1]

View File

@@ -1,51 +0,0 @@
# Bezalel Gemma 4 VPS Wiring
Issue: timmy-home #544
This helper is the repo-side operator bundle for wiring a live Gemma 4 endpoint into Bezalel's VPS config without hardcoding one dead pod forever.
What `scripts/bezalel_gemma4_vps.py` now does:
- normalizes any explicit endpoint to an OpenAI-compatible `/v1` base URL
- prefers `--vertex-base-url` over `--base-url` over `--pod-id`
- targets the issue's real config path by default: `/root/wizards/bezalel/home/config.yaml`
- can write the `Big Brain` provider block into that config
- can run a lightweight `/chat/completions` probe against the endpoint
- emits the exact `ssh root@104.131.15.18 ... curl ...` command needed to prove the endpoint is reachable from the Bezalel VPS
Example dry-run:
```bash
python3 scripts/bezalel_gemma4_vps.py \
--base-url https://<pod-id>-11434.proxy.runpod.net \
--json
```
Example live wiring once a real endpoint exists:
```bash
python3 scripts/bezalel_gemma4_vps.py \
--base-url https://<pod-id>-11434.proxy.runpod.net \
--config-path /root/wizards/bezalel/home/config.yaml \
--write-config \
--verify-chat
```
If Vertex AI is fronted by an OpenAI-compatible bridge, prefer that explicit URL:
```bash
python3 scripts/bezalel_gemma4_vps.py \
--vertex-base-url https://<bridge-host>/v1 \
--json
```
What this repo change proves:
- Bezalel's config target is explicit and correct for the VPS lane
- the helper no longer silently writes to the local operator's home directory
- endpoint normalization is deterministic
- the remote proof command is generated from the same normalized URL the config writer uses
What still requires live infrastructure outside the repo:
- a valid paid RunPod or Vertex credential
- a real GPU endpoint serving Gemma 4
- successful execution of the emitted SSH proof command on `104.131.15.18`
- successful Bezalel Hermes chat against that live endpoint

View File

@@ -8,14 +8,12 @@ Safe by default:
- can call the RunPod GraphQL API if a key is provided and --apply-runpod is used
- can update a Hermes config file in-place when --write-config is used
- can verify an OpenAI-compatible endpoint with a lightweight chat probe
- emits the exact Bezalel VPS curl proof command for remote verification
"""
from __future__ import annotations
import argparse
import json
import shlex
from pathlib import Path
from typing import Any
from urllib import request
@@ -29,9 +27,7 @@ DEFAULT_IMAGE = "ollama/ollama:latest"
DEFAULT_MODEL = "gemma4:latest"
DEFAULT_PROVIDER_NAME = "Big Brain"
DEFAULT_TOKEN_FILE = Path.home() / ".config" / "runpod" / "access_key"
DEFAULT_CONFIG_PATH = Path("/root/wizards/bezalel/home/config.yaml")
DEFAULT_BEZALEL_VPS_HOST = "104.131.15.18"
DEFAULT_VERIFY_PROMPT = "Say READY"
DEFAULT_CONFIG_PATH = Path.home() / "wizards" / "bezalel" / "home" / "config.yaml"
def build_deploy_mutation(
@@ -67,31 +63,8 @@ mutation {{
'''.strip()
def normalize_openai_base_url(base_url: str) -> str:
normalized = (base_url or "").strip().rstrip("/")
if not normalized:
return normalized
for suffix in ("/chat/completions", "/models"):
if normalized.endswith(suffix):
normalized = normalized[: -len(suffix)]
break
if not normalized.endswith("/v1"):
normalized = f"{normalized}/v1"
return normalized
def build_runpod_endpoint(pod_id: str, port: int = 11434) -> str:
return normalize_openai_base_url(f"https://{pod_id}-{port}.proxy.runpod.net")
def resolve_base_url(*, vertex_base_url: str | None = None, base_url: str | None = None, pod_id: str | None = None) -> tuple[str | None, str | None]:
if vertex_base_url:
return normalize_openai_base_url(vertex_base_url), "vertex_base_url"
if base_url:
return normalize_openai_base_url(base_url), "base_url"
if pod_id:
return build_runpod_endpoint(pod_id), "pod_id"
return None, None
return f"https://{pod_id}-{port}.proxy.runpod.net/v1"
def parse_deploy_response(payload: dict[str, Any]) -> dict[str, str]:
@@ -129,7 +102,7 @@ def update_config_text(config_text: str, *, base_url: str, model: str = DEFAULT_
replacement = {
"name": provider_name,
"base_url": normalize_openai_base_url(base_url),
"base_url": base_url,
"api_key": "",
"model": model,
}
@@ -156,8 +129,7 @@ def write_config_file(config_path: Path, *, base_url: str, model: str = DEFAULT_
return updated
def verify_openai_chat(base_url: str, *, model: str = DEFAULT_MODEL, prompt: str = DEFAULT_VERIFY_PROMPT) -> str:
base_url = normalize_openai_base_url(base_url)
def verify_openai_chat(base_url: str, *, model: str = DEFAULT_MODEL, prompt: str = "Say READY") -> str:
payload = json.dumps(
{
"model": model,
@@ -167,7 +139,7 @@ def verify_openai_chat(base_url: str, *, model: str = DEFAULT_MODEL, prompt: str
}
).encode()
req = request.Request(
f"{base_url}/chat/completions",
f"{base_url.rstrip('/')}/chat/completions",
data=payload,
headers={"Content-Type": "application/json"},
method="POST",
@@ -177,30 +149,6 @@ def verify_openai_chat(base_url: str, *, model: str = DEFAULT_MODEL, prompt: str
return data["choices"][0]["message"]["content"]
def build_vps_verify_command(
*,
base_url: str,
model: str = DEFAULT_MODEL,
prompt: str = DEFAULT_VERIFY_PROMPT,
vps_host: str = DEFAULT_BEZALEL_VPS_HOST,
) -> str:
payload = json.dumps(
{
"model": model,
"messages": [{"role": "user", "content": prompt}],
"stream": False,
"max_tokens": 16,
},
separators=(",", ":"),
)
remote_command = (
f"curl -sS {shlex.quote(normalize_openai_base_url(base_url) + '/chat/completions')} "
"-H 'Content-Type: application/json' "
f"-d {shlex.quote(payload)}"
)
return f"ssh root@{vps_host} {shlex.quote(remote_command)}"
def parse_args() -> argparse.Namespace:
parser = argparse.ArgumentParser(description="Provision a RunPod Gemma 4 endpoint and wire a Hermes config for Bezalel.")
parser.add_argument("--pod-name", default="bezalel-gemma4")
@@ -212,8 +160,6 @@ def parse_args() -> argparse.Namespace:
parser.add_argument("--config-path", type=Path, default=DEFAULT_CONFIG_PATH)
parser.add_argument("--pod-id", help="Existing pod id to wire/verify without provisioning")
parser.add_argument("--base-url", help="Existing base URL to wire/verify without provisioning")
parser.add_argument("--vertex-base-url", help="OpenAI-compatible Vertex bridge URL; takes precedence over --base-url and --pod-id")
parser.add_argument("--vps-host", default=DEFAULT_BEZALEL_VPS_HOST, help="Bezalel VPS host for the remote curl proof command")
parser.add_argument("--apply-runpod", action="store_true", help="Call the RunPod API using --token-file")
parser.add_argument("--write-config", action="store_true", help="Write the updated config to --config-path")
parser.add_argument("--verify-chat", action="store_true", help="Call the OpenAI-compatible chat endpoint")
@@ -229,18 +175,13 @@ def main() -> None:
"cloud_type": args.cloud_type,
"model": args.model,
"provider_name": args.provider_name,
"config_path": str(args.config_path),
"vps_host": args.vps_host,
"actions": [],
}
base_url, base_url_source = resolve_base_url(
vertex_base_url=args.vertex_base_url,
base_url=args.base_url,
pod_id=args.pod_id,
)
if base_url_source:
summary["actions"].append(f"resolved_base_url_from_{base_url_source}")
base_url = args.base_url
if not base_url and args.pod_id:
base_url = build_runpod_endpoint(args.pod_id)
summary["actions"].append("computed_base_url_from_pod_id")
if args.apply_runpod:
if not args.token_file.exists():
@@ -255,17 +196,12 @@ def main() -> None:
base_url = build_runpod_endpoint("<pod-id>")
summary["actions"].append("using_placeholder_base_url")
summary["base_url"] = normalize_openai_base_url(base_url)
summary["base_url"] = base_url
summary["config_preview"] = update_config_text("", base_url=base_url, model=args.model, provider_name=args.provider_name)
summary["vps_verify_command"] = build_vps_verify_command(
base_url=base_url,
model=args.model,
prompt=DEFAULT_VERIFY_PROMPT,
vps_host=args.vps_host,
)
if args.write_config:
write_config_file(args.config_path, base_url=base_url, model=args.model, provider_name=args.provider_name)
summary["config_path"] = str(args.config_path)
summary["actions"].append("wrote_config")
if args.verify_chat:
@@ -278,10 +214,8 @@ def main() -> None:
print("--- Bezalel Gemma4 RunPod Wiring ---")
print(f"Pod name: {args.pod_name}")
print(f"Base URL: {summary['base_url']}")
print(f"Base URL: {base_url}")
print(f"Model: {args.model}")
print(f"Config target: {args.config_path}")
print(f"Bezalel VPS proof: {summary['vps_verify_command']}")
if args.write_config:
print(f"Config written: {args.config_path}")
if "verify_response" in summary:

View File

@@ -1,20 +1,14 @@
from __future__ import annotations
import json
from pathlib import Path
from unittest.mock import patch
import yaml
from scripts.bezalel_gemma4_vps import (
DEFAULT_CONFIG_PATH,
DEFAULT_BEZALEL_VPS_HOST,
build_deploy_mutation,
build_runpod_endpoint,
build_vps_verify_command,
normalize_openai_base_url,
parse_deploy_response,
resolve_base_url,
update_config_text,
verify_openai_chat,
)
@@ -34,10 +28,6 @@ class _FakeResponse:
return False
def test_default_config_path_targets_bezalel_vps_root_config() -> None:
assert DEFAULT_CONFIG_PATH == Path("/root/wizards/bezalel/home/config.yaml")
def test_build_deploy_mutation_uses_ollama_image_and_openai_port() -> None:
query = build_deploy_mutation(name="bezalel-gemma4", gpu_type="NVIDIA L40S", model_tag="gemma4:latest")
@@ -47,30 +37,6 @@ def test_build_deploy_mutation_uses_ollama_image_and_openai_port() -> None:
assert 'volumeMountPath: "/root/.ollama"' in query
def test_normalize_openai_base_url_adds_v1_suffix() -> None:
assert normalize_openai_base_url("https://pod-11434.proxy.runpod.net") == "https://pod-11434.proxy.runpod.net/v1"
def test_normalize_openai_base_url_trims_chat_completions_suffix() -> None:
assert normalize_openai_base_url("https://pod-11434.proxy.runpod.net/v1/chat/completions") == "https://pod-11434.proxy.runpod.net/v1"
def test_resolve_base_url_prefers_vertex_over_base_and_pod_id() -> None:
base_url, source = resolve_base_url(
vertex_base_url="https://vertex.example.com/openai",
base_url="https://plain.example.com",
pod_id="abc123",
)
assert source == "vertex_base_url"
assert base_url == "https://vertex.example.com/openai/v1"
def test_resolve_base_url_falls_back_to_base_url_before_pod_id() -> None:
base_url, source = resolve_base_url(base_url="https://plain.example.com", pod_id="abc123")
assert source == "base_url"
assert base_url == "https://plain.example.com/v1"
def test_build_runpod_endpoint_appends_v1_suffix() -> None:
assert build_runpod_endpoint("abc123") == "https://abc123-11434.proxy.runpod.net/v1"
@@ -94,7 +60,7 @@ def test_parse_deploy_response_extracts_pod_id_and_endpoint() -> None:
}
def test_update_config_text_upserts_big_brain_provider_and_normalizes_base_url() -> None:
def test_update_config_text_upserts_big_brain_provider() -> None:
original = """
model:
default: kimi-k2.5
@@ -106,7 +72,7 @@ custom_providers:
model: gemma3:27b
"""
updated = update_config_text(original, base_url="https://new-pod-11434.proxy.runpod.net", model="gemma4:latest")
updated = update_config_text(original, base_url="https://new-pod-11434.proxy.runpod.net/v1", model="gemma4:latest")
parsed = yaml.safe_load(updated)
assert parsed["model"] == {"default": "kimi-k2.5", "provider": "kimi-coding"}
@@ -120,14 +86,7 @@ custom_providers:
]
def test_build_vps_verify_command_targets_bezalel_host_and_chat_completions() -> None:
command = build_vps_verify_command(base_url="https://pod-11434.proxy.runpod.net", model="gemma4:latest")
assert command.startswith(f"ssh root@{DEFAULT_BEZALEL_VPS_HOST} ")
assert "/v1/chat/completions" in command
assert "gemma4:latest" in command
def test_verify_openai_chat_calls_chat_completions_with_normalized_base_url() -> None:
def test_verify_openai_chat_calls_chat_completions() -> None:
response_payload = {
"choices": [
{
@@ -142,7 +101,7 @@ def test_verify_openai_chat_calls_chat_completions_with_normalized_base_url() ->
"scripts.bezalel_gemma4_vps.request.urlopen",
return_value=_FakeResponse(response_payload),
) as mocked:
result = verify_openai_chat("https://pod-11434.proxy.runpod.net", model="gemma4:latest", prompt="say READY")
result = verify_openai_chat("https://pod-11434.proxy.runpod.net/v1", model="gemma4:latest", prompt="say READY")
assert result == "READY"
req = mocked.call_args.args[0]
@@ -150,10 +109,3 @@ def test_verify_openai_chat_calls_chat_completions_with_normalized_base_url() ->
payload = json.loads(req.data.decode())
assert payload["model"] == "gemma4:latest"
assert payload["messages"][0]["content"] == "say READY"
def test_readme_documents_root_config_path_and_vps_proof_command() -> None:
readme = Path("scripts/README_bezalel_gemma4_vps.md").read_text()
assert "/root/wizards/bezalel/home/config.yaml" in readme
assert "ssh root@104.131.15.18" in readme
assert "--vertex-base-url" in readme

View File

@@ -0,0 +1,58 @@
from importlib.util import module_from_spec, spec_from_file_location
from pathlib import Path
import unittest
ROOT = Path(__file__).resolve().parent.parent
GAME_PATH = ROOT / "evennia" / "timmy_world" / "game.py"
def load_game_module():
spec = spec_from_file_location("tower_game_items", GAME_PATH)
module = module_from_spec(spec)
assert spec.loader is not None
spec.loader.exec_module(module)
module.random.seed(0)
return module
class TestTowerGameWorldItems(unittest.TestCase):
def test_world_has_ten_unique_items_and_a_quest_item(self):
module = load_game_module()
world = module.World()
room_items = {
item
for room in world.rooms.values()
for item in room.get("discoverables", [])
}
self.assertGreaterEqual(len(room_items), 10)
self.assertIn("foraged key", room_items)
self.assertTrue(module.WORLD_ITEMS["foraged key"]["quest_item"])
def test_items_change_world_state_when_used(self):
module = load_game_module()
engine = module.GameEngine()
engine.start_new_game()
engine.world.characters["Timmy"]["energy"] = 10
engine.world.characters["Timmy"]["room"] = "Garden"
initial_growth = engine.world.rooms["Garden"]["growth"]
engine.run_tick("take:seed packet")
use_seed = engine.run_tick("use:seed packet")
self.assertGreater(engine.world.rooms["Garden"]["growth"], initial_growth)
self.assertNotIn("seed packet", engine.world.characters["Timmy"]["inventory"])
self.assertTrue(any("garden" in line.lower() for line in use_seed["world_events"] + use_seed["log"]))
engine.world.characters["Timmy"]["energy"] = 10
engine.run_tick("take:foraged key")
use_key = engine.run_tick("use:foraged key")
self.assertTrue(engine.world.state["tower_cache_unlocked"])
self.assertTrue(any("cache" in line.lower() or "quest" in line.lower() for line in use_key["world_events"] + use_key["log"]))
if __name__ == "__main__":
unittest.main()