Compare commits
1 Commits
burn/667-1
...
fix/562
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
9312e4dbee |
97
.gitea/workflows/agent-pr-gate.yml
Normal file
97
.gitea/workflows/agent-pr-gate.yml
Normal file
@@ -0,0 +1,97 @@
|
||||
name: Agent PR Gate
|
||||
'on':
|
||||
pull_request:
|
||||
branches: [main]
|
||||
|
||||
jobs:
|
||||
gate:
|
||||
runs-on: ubuntu-latest
|
||||
outputs:
|
||||
syntax_status: ${{ steps.syntax.outcome }}
|
||||
tests_status: ${{ steps.tests.outcome }}
|
||||
criteria_status: ${{ steps.criteria.outcome }}
|
||||
risk_level: ${{ steps.risk.outputs.level }}
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
fetch-depth: 0
|
||||
|
||||
- uses: actions/setup-python@v5
|
||||
with:
|
||||
python-version: '3.11'
|
||||
|
||||
- name: Install CI dependencies
|
||||
run: |
|
||||
python3 -m pip install --quiet pyyaml pytest
|
||||
|
||||
- id: risk
|
||||
name: Classify PR risk
|
||||
run: |
|
||||
BASE_REF="${GITHUB_BASE_REF:-main}"
|
||||
git fetch origin "$BASE_REF" --depth 1
|
||||
git diff --name-only "origin/$BASE_REF"...HEAD > /tmp/changed_files.txt
|
||||
python3 scripts/agent_pr_gate.py classify-risk --files-file /tmp/changed_files.txt > /tmp/risk.json
|
||||
python3 - <<'PY'
|
||||
import json, os
|
||||
with open('/tmp/risk.json', 'r', encoding='utf-8') as fh:
|
||||
data = json.load(fh)
|
||||
with open(os.environ['GITHUB_OUTPUT'], 'a', encoding='utf-8') as fh:
|
||||
fh.write('level=' + data['risk'] + '\n')
|
||||
PY
|
||||
|
||||
- id: syntax
|
||||
name: Syntax and parse checks
|
||||
continue-on-error: true
|
||||
run: |
|
||||
find . \( -name '*.yml' -o -name '*.yaml' \) | grep -v .gitea | xargs -r python3 -c "import sys,yaml; [yaml.safe_load(open(f)) for f in sys.argv[1:]]"
|
||||
find . -name '*.json' | while read f; do python3 -m json.tool "$f" > /dev/null || exit 1; done
|
||||
find . -name '*.py' | xargs -r python3 -m py_compile
|
||||
find . -name '*.sh' | xargs -r bash -n
|
||||
|
||||
- id: tests
|
||||
name: Test suite
|
||||
continue-on-error: true
|
||||
run: |
|
||||
pytest -q --ignore=uni-wizard/v2/tests/test_author_whitelist.py
|
||||
|
||||
- id: criteria
|
||||
name: PR criteria verification
|
||||
continue-on-error: true
|
||||
run: |
|
||||
python3 scripts/agent_pr_gate.py validate-pr --event-path "$GITHUB_EVENT_PATH"
|
||||
|
||||
- name: Fail gate if any required check failed
|
||||
if: steps.syntax.outcome != 'success' || steps.tests.outcome != 'success' || steps.criteria.outcome != 'success'
|
||||
run: exit 1
|
||||
|
||||
report:
|
||||
needs: gate
|
||||
if: always()
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
|
||||
- uses: actions/setup-python@v5
|
||||
with:
|
||||
python-version: '3.11'
|
||||
|
||||
- name: Post PR gate report
|
||||
env:
|
||||
GITEA_TOKEN: ${{ github.token }}
|
||||
run: |
|
||||
python3 scripts/agent_pr_gate.py comment \
|
||||
--event-path "$GITHUB_EVENT_PATH" \
|
||||
--token "$GITEA_TOKEN" \
|
||||
--syntax "${{ needs.gate.outputs.syntax_status }}" \
|
||||
--tests "${{ needs.gate.outputs.tests_status }}" \
|
||||
--criteria "${{ needs.gate.outputs.criteria_status }}" \
|
||||
--risk "${{ needs.gate.outputs.risk_level }}"
|
||||
|
||||
- name: Auto-merge low-risk clean PRs
|
||||
if: needs.gate.result == 'success' && needs.gate.outputs.risk_level == 'low'
|
||||
env:
|
||||
GITEA_TOKEN: ${{ github.token }}
|
||||
run: |
|
||||
python3 scripts/agent_pr_gate.py merge \
|
||||
--event-path "$GITHUB_EVENT_PATH" \
|
||||
--token "$GITEA_TOKEN"
|
||||
@@ -1,5 +1,5 @@
|
||||
name: Smoke Test
|
||||
on:
|
||||
'on':
|
||||
pull_request:
|
||||
push:
|
||||
branches: [main]
|
||||
@@ -11,10 +11,13 @@ jobs:
|
||||
- uses: actions/setup-python@v5
|
||||
with:
|
||||
python-version: '3.11'
|
||||
- name: Install parse dependencies
|
||||
run: |
|
||||
python3 -m pip install --quiet pyyaml
|
||||
- name: Parse check
|
||||
run: |
|
||||
find . -name '*.yml' -o -name '*.yaml' | grep -v .gitea | xargs -r python3 -c "import sys,yaml; [yaml.safe_load(open(f)) for f in sys.argv[1:]]"
|
||||
find . -name '*.json' | xargs -r python3 -m json.tool > /dev/null
|
||||
find . \( -name '*.yml' -o -name '*.yaml' \) | grep -v .gitea | xargs -r python3 -c "import sys,yaml; [yaml.safe_load(open(f)) for f in sys.argv[1:]]"
|
||||
find . -name '*.json' | while read f; do python3 -m json.tool "$f" > /dev/null || exit 1; done
|
||||
find . -name '*.py' | xargs -r python3 -m py_compile
|
||||
find . -name '*.sh' | xargs -r bash -n
|
||||
echo "PASS: All files parse"
|
||||
|
||||
191
scripts/agent_pr_gate.py
Executable file
191
scripts/agent_pr_gate.py
Executable file
@@ -0,0 +1,191 @@
|
||||
#!/usr/bin/env python3
|
||||
import argparse
|
||||
import json
|
||||
import os
|
||||
import re
|
||||
import sys
|
||||
import urllib.request
|
||||
from pathlib import Path
|
||||
|
||||
API_BASE = "https://forge.alexanderwhitestone.com/api/v1"
|
||||
LOW_RISK_PREFIXES = (
|
||||
'docs/', 'reports/', 'notes/', 'tickets/', 'research/', 'briefings/',
|
||||
'twitter-archive/notes/', 'tests/'
|
||||
)
|
||||
LOW_RISK_SUFFIXES = {'.md', '.txt', '.jsonl'}
|
||||
MEDIUM_RISK_PREFIXES = ('.gitea/workflows/',)
|
||||
HIGH_RISK_PREFIXES = (
|
||||
'scripts/', 'deploy/', 'infrastructure/', 'metrics/', 'heartbeat/',
|
||||
'wizards/', 'evennia/', 'uniwizard/', 'uni-wizard/', 'timmy-local/',
|
||||
'evolution/'
|
||||
)
|
||||
HIGH_RISK_SUFFIXES = {'.py', '.sh', '.ini', '.service'}
|
||||
|
||||
|
||||
def read_changed_files(path):
|
||||
return [line.strip() for line in Path(path).read_text(encoding='utf-8').splitlines() if line.strip()]
|
||||
|
||||
|
||||
def classify_risk(files):
|
||||
if not files:
|
||||
return 'high'
|
||||
level = 'low'
|
||||
for file_path in files:
|
||||
path = file_path.strip()
|
||||
suffix = Path(path).suffix.lower()
|
||||
if path.startswith(LOW_RISK_PREFIXES):
|
||||
continue
|
||||
if path.startswith(HIGH_RISK_PREFIXES) or suffix in HIGH_RISK_SUFFIXES:
|
||||
return 'high'
|
||||
if path.startswith(MEDIUM_RISK_PREFIXES):
|
||||
level = 'medium'
|
||||
continue
|
||||
if path.startswith(LOW_RISK_PREFIXES) or suffix in LOW_RISK_SUFFIXES:
|
||||
continue
|
||||
level = 'high'
|
||||
return level
|
||||
|
||||
|
||||
def validate_pr_body(title, body):
|
||||
details = []
|
||||
combined = f"{title}\n{body}".strip()
|
||||
if not re.search(r'#\d+', combined):
|
||||
details.append('PR body/title must include an issue reference like #562.')
|
||||
if not re.search(r'(^|\n)\s*(verification|tests?)\s*:', body, re.IGNORECASE):
|
||||
details.append('PR body must include a Verification: section.')
|
||||
return (len(details) == 0, details)
|
||||
|
||||
|
||||
def build_comment_body(syntax_status, tests_status, criteria_status, risk_level):
|
||||
statuses = {
|
||||
'syntax': syntax_status,
|
||||
'tests': tests_status,
|
||||
'criteria': criteria_status,
|
||||
}
|
||||
all_clean = all(value == 'success' for value in statuses.values())
|
||||
action = 'auto-merge' if all_clean and risk_level == 'low' else 'human review'
|
||||
lines = [
|
||||
'## Agent PR Gate',
|
||||
'',
|
||||
'| Check | Status |',
|
||||
'|-------|--------|',
|
||||
f"| Syntax / parse | {syntax_status} |",
|
||||
f"| Test suite | {tests_status} |",
|
||||
f"| PR criteria | {criteria_status} |",
|
||||
f"| Risk level | {risk_level} |",
|
||||
'',
|
||||
]
|
||||
failed = [name for name, value in statuses.items() if value != 'success']
|
||||
if failed:
|
||||
lines.append('### Failure details')
|
||||
for name in failed:
|
||||
lines.append(f'- {name} reported failure. Inspect the workflow logs for that step.')
|
||||
else:
|
||||
lines.append('All automated checks passed.')
|
||||
lines.extend([
|
||||
'',
|
||||
f'Recommendation: {action}.',
|
||||
'Low-risk documentation/test-only PRs may be auto-merged. Operational changes stay in human review.',
|
||||
])
|
||||
return '\n'.join(lines)
|
||||
|
||||
|
||||
def _read_event(event_path):
|
||||
data = json.loads(Path(event_path).read_text(encoding='utf-8'))
|
||||
pr = data.get('pull_request') or {}
|
||||
repo = (data.get('repository') or {}).get('full_name') or os.environ.get('GITHUB_REPOSITORY')
|
||||
pr_number = pr.get('number') or data.get('number')
|
||||
title = pr.get('title') or ''
|
||||
body = pr.get('body') or ''
|
||||
return repo, pr_number, title, body
|
||||
|
||||
|
||||
def _request_json(method, url, token, payload=None):
|
||||
data = None if payload is None else json.dumps(payload).encode('utf-8')
|
||||
headers = {'Authorization': f'token {token}', 'Content-Type': 'application/json'}
|
||||
req = urllib.request.Request(url, data=data, headers=headers, method=method)
|
||||
with urllib.request.urlopen(req, timeout=30) as resp:
|
||||
return json.loads(resp.read().decode('utf-8'))
|
||||
|
||||
|
||||
def post_comment(repo, pr_number, token, body):
|
||||
url = f'{API_BASE}/repos/{repo}/issues/{pr_number}/comments'
|
||||
return _request_json('POST', url, token, {'body': body})
|
||||
|
||||
|
||||
def merge_pr(repo, pr_number, token):
|
||||
url = f'{API_BASE}/repos/{repo}/pulls/{pr_number}/merge'
|
||||
return _request_json('POST', url, token, {'Do': 'merge'})
|
||||
|
||||
|
||||
def cmd_classify_risk(args):
|
||||
files = list(args.files or [])
|
||||
if args.files_file:
|
||||
files.extend(read_changed_files(args.files_file))
|
||||
print(json.dumps({'risk': classify_risk(files), 'files': files}, indent=2))
|
||||
return 0
|
||||
|
||||
|
||||
def cmd_validate_pr(args):
|
||||
_, _, title, body = _read_event(args.event_path)
|
||||
ok, details = validate_pr_body(title, body)
|
||||
if ok:
|
||||
print('PR body validation passed.')
|
||||
return 0
|
||||
for detail in details:
|
||||
print(detail)
|
||||
return 1
|
||||
|
||||
|
||||
def cmd_comment(args):
|
||||
repo, pr_number, _, _ = _read_event(args.event_path)
|
||||
body = build_comment_body(args.syntax, args.tests, args.criteria, args.risk)
|
||||
post_comment(repo, pr_number, args.token, body)
|
||||
print(f'Commented on PR #{pr_number} in {repo}.')
|
||||
return 0
|
||||
|
||||
|
||||
def cmd_merge(args):
|
||||
repo, pr_number, _, _ = _read_event(args.event_path)
|
||||
merge_pr(repo, pr_number, args.token)
|
||||
print(f'Merged PR #{pr_number} in {repo}.')
|
||||
return 0
|
||||
|
||||
|
||||
def build_parser():
|
||||
parser = argparse.ArgumentParser(description='Agent PR CI helpers for timmy-home.')
|
||||
sub = parser.add_subparsers(dest='command', required=True)
|
||||
|
||||
classify = sub.add_parser('classify-risk')
|
||||
classify.add_argument('--files-file')
|
||||
classify.add_argument('files', nargs='*')
|
||||
classify.set_defaults(func=cmd_classify_risk)
|
||||
|
||||
validate = sub.add_parser('validate-pr')
|
||||
validate.add_argument('--event-path', required=True)
|
||||
validate.set_defaults(func=cmd_validate_pr)
|
||||
|
||||
comment = sub.add_parser('comment')
|
||||
comment.add_argument('--event-path', required=True)
|
||||
comment.add_argument('--token', required=True)
|
||||
comment.add_argument('--syntax', required=True)
|
||||
comment.add_argument('--tests', required=True)
|
||||
comment.add_argument('--criteria', required=True)
|
||||
comment.add_argument('--risk', required=True)
|
||||
comment.set_defaults(func=cmd_comment)
|
||||
|
||||
merge = sub.add_parser('merge')
|
||||
merge.add_argument('--event-path', required=True)
|
||||
merge.add_argument('--token', required=True)
|
||||
merge.set_defaults(func=cmd_merge)
|
||||
return parser
|
||||
|
||||
|
||||
def main(argv=None):
|
||||
parser = build_parser()
|
||||
args = parser.parse_args(argv)
|
||||
return args.func(args)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
sys.exit(main())
|
||||
@@ -1,290 +0,0 @@
|
||||
#!/usr/bin/env python3
|
||||
"""Codebase Test Generator — Fill Coverage Gaps (#667)."""
|
||||
|
||||
import ast
|
||||
import os
|
||||
import sys
|
||||
import argparse
|
||||
from dataclasses import dataclass, field
|
||||
from pathlib import Path
|
||||
from typing import Dict, List, Optional, Set, Tuple
|
||||
|
||||
|
||||
@dataclass
|
||||
class FunctionInfo:
|
||||
name: str
|
||||
module_path: str
|
||||
class_name: Optional[str] = None
|
||||
lineno: int = 0
|
||||
args: List[str] = field(default_factory=list)
|
||||
is_async: bool = False
|
||||
is_private: bool = False
|
||||
is_property: bool = False
|
||||
docstring: Optional[str] = None
|
||||
has_return: bool = False
|
||||
raises: List[str] = field(default_factory=list)
|
||||
decorators: List[str] = field(default_factory=list)
|
||||
|
||||
@property
|
||||
def qualified_name(self):
|
||||
if self.class_name:
|
||||
return f"{self.class_name}.{self.name}"
|
||||
return self.name
|
||||
|
||||
@property
|
||||
def test_name(self):
|
||||
safe_mod = self.module_path.replace("/", "_").replace(".py", "").replace("-", "_")
|
||||
safe_cls = self.class_name + "_" if self.class_name else ""
|
||||
return f"test_{safe_mod}_{safe_cls}{self.name}"
|
||||
|
||||
|
||||
@dataclass
|
||||
class CoverageGap:
|
||||
func: FunctionInfo
|
||||
reason: str
|
||||
test_priority: int
|
||||
|
||||
|
||||
class SourceAnalyzer(ast.NodeVisitor):
|
||||
def __init__(self, module_path: str):
|
||||
self.module_path = module_path
|
||||
self.functions: List[FunctionInfo] = []
|
||||
self._class_stack: List[str] = []
|
||||
|
||||
def visit_ClassDef(self, node):
|
||||
self._class_stack.append(node.name)
|
||||
self.generic_visit(node)
|
||||
self._class_stack.pop()
|
||||
|
||||
def visit_FunctionDef(self, node):
|
||||
self._collect(node, False)
|
||||
self.generic_visit(node)
|
||||
|
||||
def visit_AsyncFunctionDef(self, node):
|
||||
self._collect(node, True)
|
||||
self.generic_visit(node)
|
||||
|
||||
def _collect(self, node, is_async):
|
||||
cls = self._class_stack[-1] if self._class_stack else None
|
||||
args = [a.arg for a in node.args.args if a.arg not in ("self", "cls")]
|
||||
has_ret = any(isinstance(c, ast.Return) and c.value for c in ast.walk(node))
|
||||
raises = []
|
||||
for c in ast.walk(node):
|
||||
if isinstance(c, ast.Raise) and c.exc:
|
||||
if isinstance(c.exc, ast.Call) and isinstance(c.exc.func, ast.Name):
|
||||
raises.append(c.exc.func.id)
|
||||
decos = []
|
||||
for d in node.decorator_list:
|
||||
if isinstance(d, ast.Name): decos.append(d.id)
|
||||
elif isinstance(d, ast.Attribute): decos.append(d.attr)
|
||||
self.functions.append(FunctionInfo(
|
||||
name=node.name, module_path=self.module_path, class_name=cls,
|
||||
lineno=node.lineno, args=args, is_async=is_async,
|
||||
is_private=node.name.startswith("_") and not node.name.startswith("__"),
|
||||
is_property="property" in decos,
|
||||
docstring=ast.get_docstring(node), has_return=has_ret,
|
||||
raises=raises, decorators=decos))
|
||||
|
||||
|
||||
def analyze_file(filepath, base_dir):
|
||||
module_path = os.path.relpath(filepath, base_dir)
|
||||
try:
|
||||
with open(filepath, "r", errors="replace") as f:
|
||||
tree = ast.parse(f.read(), filename=filepath)
|
||||
except (SyntaxError, UnicodeDecodeError):
|
||||
return []
|
||||
a = SourceAnalyzer(module_path)
|
||||
a.visit(tree)
|
||||
return a.functions
|
||||
|
||||
|
||||
def find_source_files(source_dir):
|
||||
exclude = {"__pycache__", ".git", "venv", ".venv", "node_modules", ".tox", "build", "dist"}
|
||||
files = []
|
||||
for root, dirs, fs in os.walk(source_dir):
|
||||
dirs[:] = [d for d in dirs if d not in exclude and not d.startswith(".")]
|
||||
for f in fs:
|
||||
if f.endswith(".py") and f != "__init__.py" and not f.startswith("test_"):
|
||||
files.append(os.path.join(root, f))
|
||||
return sorted(files)
|
||||
|
||||
|
||||
def find_existing_tests(test_dir):
|
||||
existing = set()
|
||||
for root, dirs, fs in os.walk(test_dir):
|
||||
for f in fs:
|
||||
if f.startswith("test_") and f.endswith(".py"):
|
||||
try:
|
||||
with open(os.path.join(root, f)) as fh:
|
||||
tree = ast.parse(fh.read())
|
||||
for node in ast.walk(tree):
|
||||
if isinstance(node, ast.FunctionDef) and node.name.startswith("test_"):
|
||||
existing.add(node.name)
|
||||
except (SyntaxError, UnicodeDecodeError):
|
||||
pass
|
||||
return existing
|
||||
|
||||
|
||||
def identify_gaps(functions, existing_tests):
|
||||
gaps = []
|
||||
for func in functions:
|
||||
if func.name.startswith("__") and func.name != "__init__":
|
||||
continue
|
||||
covered = func.name in str(existing_tests)
|
||||
if not covered:
|
||||
pri = 3 if func.is_private else (1 if (func.raises or func.has_return) else 2)
|
||||
gaps.append(CoverageGap(func=func, reason="no test found", test_priority=pri))
|
||||
gaps.sort(key=lambda g: (g.test_priority, g.func.module_path, g.func.name))
|
||||
return gaps
|
||||
|
||||
|
||||
def generate_test(gap):
|
||||
func = gap.func
|
||||
lines = []
|
||||
lines.append(f" # AUTO-GENERATED -- review before merging")
|
||||
lines.append(f" # Source: {func.module_path}:{func.lineno}")
|
||||
lines.append(f" # Function: {func.qualified_name}")
|
||||
lines.append("")
|
||||
mod_imp = func.module_path.replace("/", ".").replace("-", "_").replace(".py", "")
|
||||
|
||||
call_args = []
|
||||
for a in func.args:
|
||||
if a in ("self", "cls"): continue
|
||||
if "path" in a or "file" in a or "dir" in a: call_args.append(f"{a}='/tmp/test'")
|
||||
elif "name" in a: call_args.append(f"{a}='test'")
|
||||
elif "id" in a or "key" in a: call_args.append(f"{a}='test_id'")
|
||||
elif "message" in a or "text" in a: call_args.append(f"{a}='test msg'")
|
||||
elif "count" in a or "num" in a or "size" in a: call_args.append(f"{a}=1")
|
||||
elif "flag" in a or "enabled" in a or "verbose" in a: call_args.append(f"{a}=False")
|
||||
else: call_args.append(f"{a}=None")
|
||||
args_str = ", ".join(call_args)
|
||||
|
||||
if func.is_async:
|
||||
lines.append(" @pytest.mark.asyncio")
|
||||
lines.append(f" def {func.test_name}(self):")
|
||||
lines.append(f' """Test {func.qualified_name} -- auto-generated."""')
|
||||
|
||||
if func.class_name:
|
||||
lines.append(f" try:")
|
||||
lines.append(f" from {mod_imp} import {func.class_name}")
|
||||
if func.is_private:
|
||||
lines.append(f" pytest.skip('Private method')")
|
||||
elif func.is_property:
|
||||
lines.append(f" obj = {func.class_name}()")
|
||||
lines.append(f" _ = obj.{func.name}")
|
||||
else:
|
||||
if func.raises:
|
||||
lines.append(f" with pytest.raises(({', '.join(func.raises)})):")
|
||||
lines.append(f" {func.class_name}().{func.name}({args_str})")
|
||||
else:
|
||||
lines.append(f" obj = {func.class_name}()")
|
||||
lines.append(f" result = obj.{func.name}({args_str})")
|
||||
if func.has_return:
|
||||
lines.append(f" assert result is not None or result is None # Placeholder")
|
||||
lines.append(f" except ImportError:")
|
||||
lines.append(f" pytest.skip('Module not importable')")
|
||||
else:
|
||||
lines.append(f" try:")
|
||||
lines.append(f" from {mod_imp} import {func.name}")
|
||||
if func.is_private:
|
||||
lines.append(f" pytest.skip('Private function')")
|
||||
else:
|
||||
if func.raises:
|
||||
lines.append(f" with pytest.raises(({', '.join(func.raises)})):")
|
||||
lines.append(f" {func.name}({args_str})")
|
||||
else:
|
||||
lines.append(f" result = {func.name}({args_str})")
|
||||
if func.has_return:
|
||||
lines.append(f" assert result is not None or result is None # Placeholder")
|
||||
lines.append(f" except ImportError:")
|
||||
lines.append(f" pytest.skip('Module not importable')")
|
||||
|
||||
return chr(10).join(lines)
|
||||
|
||||
|
||||
def generate_test_suite(gaps, max_tests=50):
|
||||
by_module = {}
|
||||
for gap in gaps[:max_tests]:
|
||||
by_module.setdefault(gap.func.module_path, []).append(gap)
|
||||
|
||||
lines = []
|
||||
lines.append('"""Auto-generated test suite -- Codebase Genome (#667).')
|
||||
lines.append("")
|
||||
lines.append("Generated by scripts/codebase_test_generator.py")
|
||||
lines.append("Coverage gaps identified from AST analysis.")
|
||||
lines.append("")
|
||||
lines.append("These tests are starting points. Review before merging.")
|
||||
lines.append('"""')
|
||||
lines.append("")
|
||||
lines.append("import pytest")
|
||||
lines.append("from unittest.mock import MagicMock, patch")
|
||||
lines.append("")
|
||||
lines.append("")
|
||||
lines.append("# AUTO-GENERATED -- DO NOT EDIT WITHOUT REVIEW")
|
||||
|
||||
for module, mgaps in sorted(by_module.items()):
|
||||
safe = module.replace("/", "_").replace(".py", "").replace("-", "_")
|
||||
cls_name = "".join(w.title() for w in safe.split("_"))
|
||||
lines.append("")
|
||||
lines.append(f"class Test{cls_name}Generated:")
|
||||
lines.append(f' """Auto-generated tests for {module}."""')
|
||||
for gap in mgaps:
|
||||
lines.append("")
|
||||
lines.append(generate_test(gap))
|
||||
lines.append("")
|
||||
|
||||
return chr(10).join(lines)
|
||||
|
||||
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(description="Codebase Test Generator")
|
||||
parser.add_argument("--source", default=".")
|
||||
parser.add_argument("--output", default="tests/test_genome_generated.py")
|
||||
parser.add_argument("--max-tests", type=int, default=50)
|
||||
parser.add_argument("--dry-run", action="store_true")
|
||||
parser.add_argument("--include-private", action="store_true")
|
||||
args = parser.parse_args()
|
||||
|
||||
source_dir = os.path.abspath(args.source)
|
||||
test_dir = os.path.join(source_dir, "tests")
|
||||
|
||||
print(f"Scanning: {source_dir}")
|
||||
source_files = find_source_files(source_dir)
|
||||
print(f"Source files: {len(source_files)}")
|
||||
|
||||
all_funcs = []
|
||||
for f in source_files:
|
||||
all_funcs.extend(analyze_file(f, source_dir))
|
||||
print(f"Functions/methods: {len(all_funcs)}")
|
||||
|
||||
existing = find_existing_tests(test_dir)
|
||||
print(f"Existing tests: {len(existing)}")
|
||||
|
||||
gaps = identify_gaps(all_funcs, existing)
|
||||
if not args.include_private:
|
||||
gaps = [g for g in gaps if not g.func.is_private]
|
||||
print(f"Coverage gaps: {len(gaps)}")
|
||||
|
||||
by_pri = {1: 0, 2: 0, 3: 0}
|
||||
for g in gaps:
|
||||
by_pri[g.test_priority] += 1
|
||||
print(f" High: {by_pri[1]}, Medium: {by_pri[2]}, Low: {by_pri[3]}")
|
||||
|
||||
if args.dry_run:
|
||||
for g in gaps[:10]:
|
||||
print(f" {g.func.module_path}:{g.func.lineno} {g.func.qualified_name}")
|
||||
return
|
||||
|
||||
if gaps:
|
||||
content = generate_test_suite(gaps, max_tests=args.max-tests if hasattr(args, 'max-tests') else args.max_tests)
|
||||
out = os.path.join(source_dir, args.output)
|
||||
os.makedirs(os.path.dirname(out), exist_ok=True)
|
||||
with open(out, "w") as f:
|
||||
f.write(content)
|
||||
print(f"Generated {min(len(gaps), args.max_tests)} tests -> {args.output}")
|
||||
else:
|
||||
print("No gaps found!")
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
68
tests/test_agent_pr_gate.py
Normal file
68
tests/test_agent_pr_gate.py
Normal file
@@ -0,0 +1,68 @@
|
||||
import pathlib
|
||||
import sys
|
||||
import tempfile
|
||||
import unittest
|
||||
|
||||
ROOT = pathlib.Path(__file__).resolve().parents[1]
|
||||
sys.path.insert(0, str(ROOT / 'scripts'))
|
||||
|
||||
import agent_pr_gate # noqa: E402
|
||||
|
||||
|
||||
class TestAgentPrGate(unittest.TestCase):
|
||||
def test_classify_risk_low_for_docs_and_tests_only(self):
|
||||
level = agent_pr_gate.classify_risk([
|
||||
'docs/runbook.md',
|
||||
'reports/daily-summary.md',
|
||||
'tests/test_agent_pr_gate.py',
|
||||
])
|
||||
self.assertEqual(level, 'low')
|
||||
|
||||
def test_classify_risk_high_for_operational_paths(self):
|
||||
level = agent_pr_gate.classify_risk([
|
||||
'scripts/failover_monitor.py',
|
||||
'deploy/playbook.yml',
|
||||
])
|
||||
self.assertEqual(level, 'high')
|
||||
|
||||
def test_validate_pr_body_requires_issue_ref_and_verification(self):
|
||||
ok, details = agent_pr_gate.validate_pr_body(
|
||||
'feat: add thing',
|
||||
'What changed only\n\nNo verification section here.'
|
||||
)
|
||||
self.assertFalse(ok)
|
||||
self.assertIn('issue reference', ' '.join(details).lower())
|
||||
self.assertIn('verification', ' '.join(details).lower())
|
||||
|
||||
def test_validate_pr_body_accepts_issue_ref_and_verification(self):
|
||||
ok, details = agent_pr_gate.validate_pr_body(
|
||||
'feat: add thing (#562)',
|
||||
'Refs #562\n\nVerification:\n- pytest -q\n'
|
||||
)
|
||||
self.assertTrue(ok)
|
||||
self.assertEqual(details, [])
|
||||
|
||||
def test_build_comment_body_reports_failures_and_human_review(self):
|
||||
body = agent_pr_gate.build_comment_body(
|
||||
syntax_status='success',
|
||||
tests_status='failure',
|
||||
criteria_status='success',
|
||||
risk_level='high',
|
||||
)
|
||||
self.assertIn('tests', body.lower())
|
||||
self.assertIn('failure', body.lower())
|
||||
self.assertIn('human review', body.lower())
|
||||
|
||||
def test_changed_files_file_loader_ignores_blanks(self):
|
||||
with tempfile.NamedTemporaryFile('w+', delete=False) as handle:
|
||||
handle.write('docs/one.md\n\nreports/two.md\n')
|
||||
path = handle.name
|
||||
try:
|
||||
files = agent_pr_gate.read_changed_files(path)
|
||||
finally:
|
||||
pathlib.Path(path).unlink(missing_ok=True)
|
||||
self.assertEqual(files, ['docs/one.md', 'reports/two.md'])
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
unittest.main()
|
||||
24
tests/test_agent_pr_workflow.py
Normal file
24
tests/test_agent_pr_workflow.py
Normal file
@@ -0,0 +1,24 @@
|
||||
import pathlib
|
||||
import unittest
|
||||
import yaml
|
||||
|
||||
ROOT = pathlib.Path(__file__).resolve().parents[1]
|
||||
WORKFLOW = ROOT / '.gitea' / 'workflows' / 'agent-pr-gate.yml'
|
||||
|
||||
|
||||
class TestAgentPrWorkflow(unittest.TestCase):
|
||||
def test_workflow_exists(self):
|
||||
self.assertTrue(WORKFLOW.exists(), 'agent-pr-gate workflow should exist')
|
||||
|
||||
def test_workflow_has_pr_gate_and_reporting_jobs(self):
|
||||
data = yaml.safe_load(WORKFLOW.read_text(encoding='utf-8'))
|
||||
self.assertIn('pull_request', data.get('on', {}))
|
||||
jobs = data.get('jobs', {})
|
||||
self.assertIn('gate', jobs)
|
||||
self.assertIn('report', jobs)
|
||||
report_steps = jobs['report']['steps']
|
||||
self.assertTrue(any('Auto-merge low-risk clean PRs' in (step.get('name') or '') for step in report_steps))
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
unittest.main()
|
||||
@@ -1,737 +0,0 @@
|
||||
"""Auto-generated test suite -- Codebase Genome (#667).
|
||||
|
||||
Generated by scripts/codebase_test_generator.py
|
||||
Coverage gaps identified from AST analysis.
|
||||
|
||||
These tests are starting points. Review before merging.
|
||||
"""
|
||||
|
||||
import pytest
|
||||
from unittest.mock import MagicMock, patch
|
||||
|
||||
|
||||
# AUTO-GENERATED -- DO NOT EDIT WITHOUT REVIEW
|
||||
|
||||
class TestAngbandMcpServerGenerated:
|
||||
"""Auto-generated tests for angband/mcp_server.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:319
|
||||
# Function: call_tool
|
||||
|
||||
@pytest.mark.asyncio
|
||||
def test_angband_mcp_server_call_tool(self):
|
||||
"""Test call_tool -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import call_tool
|
||||
result = call_tool(name='test', arguments=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:64
|
||||
# Function: capture_screen
|
||||
|
||||
def test_angband_mcp_server_capture_screen(self):
|
||||
"""Test capture_screen -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import capture_screen
|
||||
result = capture_screen(lines=None, session_name='test')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:74
|
||||
# Function: has_save
|
||||
|
||||
def test_angband_mcp_server_has_save(self):
|
||||
"""Test has_save -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import has_save
|
||||
result = has_save(user=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:234
|
||||
# Function: keypress
|
||||
|
||||
def test_angband_mcp_server_keypress(self):
|
||||
"""Test keypress -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import keypress
|
||||
result = keypress(key='test_id', wait_ms=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:141
|
||||
# Function: launch_game
|
||||
|
||||
def test_angband_mcp_server_launch_game(self):
|
||||
"""Test launch_game -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import launch_game
|
||||
result = launch_game(user=None, new_game=None, continue_splash=None, width='test_id', height=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:253
|
||||
# Function: list_tools
|
||||
|
||||
@pytest.mark.asyncio
|
||||
def test_angband_mcp_server_list_tools(self):
|
||||
"""Test list_tools -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import list_tools
|
||||
result = list_tools()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:130
|
||||
# Function: maybe_continue_splash
|
||||
|
||||
def test_angband_mcp_server_maybe_continue_splash(self):
|
||||
"""Test maybe_continue_splash -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import maybe_continue_splash
|
||||
result = maybe_continue_splash(session_name='test')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:226
|
||||
# Function: observe
|
||||
|
||||
def test_angband_mcp_server_observe(self):
|
||||
"""Test observe -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import observe
|
||||
result = observe(lines=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:57
|
||||
# Function: pane_id
|
||||
|
||||
def test_angband_mcp_server_pane_id(self):
|
||||
"""Test pane_id -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import pane_id
|
||||
result = pane_id(session_name='test')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:108
|
||||
# Function: send_key
|
||||
|
||||
def test_angband_mcp_server_send_key(self):
|
||||
"""Test send_key -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import send_key
|
||||
with pytest.raises((RuntimeError)):
|
||||
send_key(key='test_id', session_name='test')
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:123
|
||||
# Function: send_text
|
||||
|
||||
def test_angband_mcp_server_send_text(self):
|
||||
"""Test send_text -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import send_text
|
||||
with pytest.raises((RuntimeError)):
|
||||
send_text(text='test msg', session_name='test')
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:53
|
||||
# Function: session_exists
|
||||
|
||||
def test_angband_mcp_server_session_exists(self):
|
||||
"""Test session_exists -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import session_exists
|
||||
result = session_exists(session_name='test')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:203
|
||||
# Function: stop_game
|
||||
|
||||
def test_angband_mcp_server_stop_game(self):
|
||||
"""Test stop_game -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import stop_game
|
||||
result = stop_game()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:46
|
||||
# Function: tmux
|
||||
|
||||
def test_angband_mcp_server_tmux(self):
|
||||
"""Test tmux -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import tmux
|
||||
with pytest.raises((RuntimeError)):
|
||||
tmux(args=None, check=None)
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: angband/mcp_server.py:243
|
||||
# Function: type_and_observe
|
||||
|
||||
def test_angband_mcp_server_type_and_observe(self):
|
||||
"""Test type_and_observe -- auto-generated."""
|
||||
try:
|
||||
from angband.mcp_server import type_and_observe
|
||||
result = type_and_observe(text='test msg', wait_ms=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestEvenniaTimmyWorldGameGenerated:
|
||||
"""Auto-generated tests for evennia/timmy_world/game.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/game.py:495
|
||||
# Function: ActionSystem.get_available_actions
|
||||
|
||||
def test_evennia_timmy_world_game_ActionSystem_get_available_actions(self):
|
||||
"""Test ActionSystem.get_available_actions -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.game import ActionSystem
|
||||
obj = ActionSystem()
|
||||
result = obj.get_available_actions(char_name='test', world=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/game.py:1485
|
||||
# Function: PlayerInterface.get_available_actions
|
||||
|
||||
def test_evennia_timmy_world_game_PlayerInterface_get_available_actions(self):
|
||||
"""Test PlayerInterface.get_available_actions -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.game import PlayerInterface
|
||||
obj = PlayerInterface()
|
||||
result = obj.get_available_actions()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/game.py:55
|
||||
# Function: get_narrative_phase
|
||||
|
||||
def test_evennia_timmy_world_game_get_narrative_phase(self):
|
||||
"""Test get_narrative_phase -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.game import get_narrative_phase
|
||||
result = get_narrative_phase(tick=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/game.py:65
|
||||
# Function: get_phase_transition_event
|
||||
|
||||
def test_evennia_timmy_world_game_get_phase_transition_event(self):
|
||||
"""Test get_phase_transition_event -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.game import get_phase_transition_event
|
||||
result = get_phase_transition_event(old_phase=None, new_phase=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/game.py:347
|
||||
# Function: World.get_room_desc
|
||||
|
||||
def test_evennia_timmy_world_game_World_get_room_desc(self):
|
||||
"""Test World.get_room_desc -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.game import World
|
||||
obj = World()
|
||||
result = obj.get_room_desc(room_name='test', char_name='test')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/game.py:1045
|
||||
# Function: GameEngine.load_game
|
||||
|
||||
def test_evennia_timmy_world_game_GameEngine_load_game(self):
|
||||
"""Test GameEngine.load_game -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.game import GameEngine
|
||||
obj = GameEngine()
|
||||
result = obj.load_game()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/game.py:556
|
||||
# Function: NPCAI.make_choice
|
||||
|
||||
def test_evennia_timmy_world_game_NPCAI_make_choice(self):
|
||||
"""Test NPCAI.make_choice -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.game import NPCAI
|
||||
obj = NPCAI()
|
||||
result = obj.make_choice(char_name='test')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/game.py:1454
|
||||
# Function: GameEngine.play_turn
|
||||
|
||||
def test_evennia_timmy_world_game_GameEngine_play_turn(self):
|
||||
"""Test GameEngine.play_turn -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.game import GameEngine
|
||||
obj = GameEngine()
|
||||
result = obj.play_turn(action=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/game.py:1076
|
||||
# Function: GameEngine.run_tick
|
||||
|
||||
def test_evennia_timmy_world_game_GameEngine_run_tick(self):
|
||||
"""Test GameEngine.run_tick -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.game import GameEngine
|
||||
obj = GameEngine()
|
||||
result = obj.run_tick(timmy_action=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestEvenniaTimmyWorldServerConfWebPluginsGenerated:
|
||||
"""Auto-generated tests for evennia/timmy_world/server/conf/web_plugins.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/server/conf/web_plugins.py:31
|
||||
# Function: at_webproxy_root_creation
|
||||
|
||||
def test_evennia_timmy_world_server_conf_web_plugins_at_webproxy_root_creation(self):
|
||||
"""Test at_webproxy_root_creation -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.server.conf.web_plugins import at_webproxy_root_creation
|
||||
result = at_webproxy_root_creation(web_root=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/server/conf/web_plugins.py:6
|
||||
# Function: at_webserver_root_creation
|
||||
|
||||
def test_evennia_timmy_world_server_conf_web_plugins_at_webserver_root_creation(self):
|
||||
"""Test at_webserver_root_creation -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.server.conf.web_plugins import at_webserver_root_creation
|
||||
result = at_webserver_root_creation(web_root=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestEvenniaTimmyWorldWorldGameGenerated:
|
||||
"""Auto-generated tests for evennia/timmy_world/world/game.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/world/game.py:400
|
||||
# Function: ActionSystem.get_available_actions
|
||||
|
||||
def test_evennia_timmy_world_world_game_ActionSystem_get_available_actions(self):
|
||||
"""Test ActionSystem.get_available_actions -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.world.game import ActionSystem
|
||||
obj = ActionSystem()
|
||||
result = obj.get_available_actions(char_name='test', world=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/world/game.py:1289
|
||||
# Function: PlayerInterface.get_available_actions
|
||||
|
||||
def test_evennia_timmy_world_world_game_PlayerInterface_get_available_actions(self):
|
||||
"""Test PlayerInterface.get_available_actions -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.world.game import PlayerInterface
|
||||
obj = PlayerInterface()
|
||||
result = obj.get_available_actions()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/world/game.py:254
|
||||
# Function: World.get_room_desc
|
||||
|
||||
def test_evennia_timmy_world_world_game_World_get_room_desc(self):
|
||||
"""Test World.get_room_desc -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.world.game import World
|
||||
obj = World()
|
||||
result = obj.get_room_desc(room_name='test', char_name='test')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/world/game.py:880
|
||||
# Function: GameEngine.load_game
|
||||
|
||||
def test_evennia_timmy_world_world_game_GameEngine_load_game(self):
|
||||
"""Test GameEngine.load_game -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.world.game import GameEngine
|
||||
obj = GameEngine()
|
||||
result = obj.load_game()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/world/game.py:461
|
||||
# Function: NPCAI.make_choice
|
||||
|
||||
def test_evennia_timmy_world_world_game_NPCAI_make_choice(self):
|
||||
"""Test NPCAI.make_choice -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.world.game import NPCAI
|
||||
obj = NPCAI()
|
||||
result = obj.make_choice(char_name='test')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/world/game.py:1258
|
||||
# Function: GameEngine.play_turn
|
||||
|
||||
def test_evennia_timmy_world_world_game_GameEngine_play_turn(self):
|
||||
"""Test GameEngine.play_turn -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.world.game import GameEngine
|
||||
obj = GameEngine()
|
||||
result = obj.play_turn(action=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/world/game.py:911
|
||||
# Function: GameEngine.run_tick
|
||||
|
||||
def test_evennia_timmy_world_world_game_GameEngine_run_tick(self):
|
||||
"""Test GameEngine.run_tick -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.world.game import GameEngine
|
||||
obj = GameEngine()
|
||||
result = obj.run_tick(timmy_action=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia/timmy_world/world/game.py:749
|
||||
# Function: DialogueSystem.select
|
||||
|
||||
def test_evennia_timmy_world_world_game_DialogueSystem_select(self):
|
||||
"""Test DialogueSystem.select -- auto-generated."""
|
||||
try:
|
||||
from evennia.timmy_world.world.game import DialogueSystem
|
||||
obj = DialogueSystem()
|
||||
result = obj.select(char_name='test', listener=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestEvenniaToolsLayoutGenerated:
|
||||
"""Auto-generated tests for evennia_tools/layout.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia_tools/layout.py:58
|
||||
# Function: grouped_exits
|
||||
|
||||
def test_evennia_tools_layout_grouped_exits(self):
|
||||
"""Test grouped_exits -- auto-generated."""
|
||||
try:
|
||||
from evennia_tools.layout import grouped_exits
|
||||
result = grouped_exits()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia_tools/layout.py:54
|
||||
# Function: room_keys
|
||||
|
||||
def test_evennia_tools_layout_room_keys(self):
|
||||
"""Test room_keys -- auto-generated."""
|
||||
try:
|
||||
from evennia_tools.layout import room_keys
|
||||
result = room_keys()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestEvenniaToolsTelemetryGenerated:
|
||||
"""Auto-generated tests for evennia_tools/telemetry.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia_tools/telemetry.py:8
|
||||
# Function: telemetry_dir
|
||||
|
||||
def test_evennia_tools_telemetry_telemetry_dir(self):
|
||||
"""Test telemetry_dir -- auto-generated."""
|
||||
try:
|
||||
from evennia_tools.telemetry import telemetry_dir
|
||||
result = telemetry_dir(base_dir='/tmp/test')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestEvenniaToolsTrainingGenerated:
|
||||
"""Auto-generated tests for evennia_tools/training.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia_tools/training.py:18
|
||||
# Function: example_eval_path
|
||||
|
||||
def test_evennia_tools_training_example_eval_path(self):
|
||||
"""Test example_eval_path -- auto-generated."""
|
||||
try:
|
||||
from evennia_tools.training import example_eval_path
|
||||
result = example_eval_path(repo_root=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evennia_tools/training.py:14
|
||||
# Function: example_trace_path
|
||||
|
||||
def test_evennia_tools_training_example_trace_path(self):
|
||||
"""Test example_trace_path -- auto-generated."""
|
||||
try:
|
||||
from evennia_tools.training import example_trace_path
|
||||
result = example_trace_path(repo_root=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestEvolutionBitcoinScripterGenerated:
|
||||
"""Auto-generated tests for evolution/bitcoin_scripter.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evolution/bitcoin_scripter.py:18
|
||||
# Function: BitcoinScripter.generate_script
|
||||
|
||||
def test_evolution_bitcoin_scripter_BitcoinScripter_generate_script(self):
|
||||
"""Test BitcoinScripter.generate_script -- auto-generated."""
|
||||
try:
|
||||
from evolution.bitcoin_scripter import BitcoinScripter
|
||||
obj = BitcoinScripter()
|
||||
result = obj.generate_script(requirements=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestEvolutionLightningClientGenerated:
|
||||
"""Auto-generated tests for evolution/lightning_client.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evolution/lightning_client.py:18
|
||||
# Function: LightningClient.plan_payment_route
|
||||
|
||||
def test_evolution_lightning_client_LightningClient_plan_payment_route(self):
|
||||
"""Test LightningClient.plan_payment_route -- auto-generated."""
|
||||
try:
|
||||
from evolution.lightning_client import LightningClient
|
||||
obj = LightningClient()
|
||||
result = obj.plan_payment_route(destination=None, amount_sats=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestEvolutionSovereignAccountantGenerated:
|
||||
"""Auto-generated tests for evolution/sovereign_accountant.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: evolution/sovereign_accountant.py:17
|
||||
# Function: SovereignAccountant.generate_financial_report
|
||||
|
||||
def test_evolution_sovereign_accountant_SovereignAccountant_generate_financial_report(self):
|
||||
"""Test SovereignAccountant.generate_financial_report -- auto-generated."""
|
||||
try:
|
||||
from evolution.sovereign_accountant import SovereignAccountant
|
||||
obj = SovereignAccountant()
|
||||
result = obj.generate_financial_report(transaction_history=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestInfrastructureTimmyBridgeClientTimmyClientGenerated:
|
||||
"""Auto-generated tests for infrastructure/timmy-bridge/client/timmy_client.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: infrastructure/timmy-bridge/client/timmy_client.py:108
|
||||
# Function: TimmyClient.create_artifact
|
||||
|
||||
def test_infrastructure_timmy_bridge_client_timmy_client_TimmyClient_create_artifact(self):
|
||||
"""Test TimmyClient.create_artifact -- auto-generated."""
|
||||
try:
|
||||
from infrastructure.timmy_bridge.client.timmy_client import TimmyClient
|
||||
obj = TimmyClient()
|
||||
result = obj.create_artifact()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: infrastructure/timmy-bridge/client/timmy_client.py:167
|
||||
# Function: TimmyClient.create_event
|
||||
|
||||
def test_infrastructure_timmy_bridge_client_timmy_client_TimmyClient_create_event(self):
|
||||
"""Test TimmyClient.create_event -- auto-generated."""
|
||||
try:
|
||||
from infrastructure.timmy_bridge.client.timmy_client import TimmyClient
|
||||
obj = TimmyClient()
|
||||
result = obj.create_event(kind=None, content=None, tags=None)
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: infrastructure/timmy-bridge/client/timmy_client.py:74
|
||||
# Function: TimmyClient.generate_observation
|
||||
|
||||
def test_infrastructure_timmy_bridge_client_timmy_client_TimmyClient_generate_observation(self):
|
||||
"""Test TimmyClient.generate_observation -- auto-generated."""
|
||||
try:
|
||||
from infrastructure.timmy_bridge.client.timmy_client import TimmyClient
|
||||
obj = TimmyClient()
|
||||
result = obj.generate_observation()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
|
||||
class TestInfrastructureTimmyBridgeMlxMlxIntegrationGenerated:
|
||||
"""Auto-generated tests for infrastructure/timmy-bridge/mlx/mlx_integration.py."""
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: infrastructure/timmy-bridge/mlx/mlx_integration.py:122
|
||||
# Function: MLXInference.available
|
||||
|
||||
def test_infrastructure_timmy_bridge_mlx_mlx_integration_MLXInference_available(self):
|
||||
"""Test MLXInference.available -- auto-generated."""
|
||||
try:
|
||||
from infrastructure.timmy_bridge.mlx.mlx_integration import MLXInference
|
||||
obj = MLXInference()
|
||||
_ = obj.available
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: infrastructure/timmy-bridge/mlx/mlx_integration.py:125
|
||||
# Function: MLXInference.get_stats
|
||||
|
||||
def test_infrastructure_timmy_bridge_mlx_mlx_integration_MLXInference_get_stats(self):
|
||||
"""Test MLXInference.get_stats -- auto-generated."""
|
||||
try:
|
||||
from infrastructure.timmy_bridge.mlx.mlx_integration import MLXInference
|
||||
obj = MLXInference()
|
||||
result = obj.get_stats()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: infrastructure/timmy-bridge/mlx/mlx_integration.py:30
|
||||
# Function: MLXInference.load_model
|
||||
|
||||
def test_infrastructure_timmy_bridge_mlx_mlx_integration_MLXInference_load_model(self):
|
||||
"""Test MLXInference.load_model -- auto-generated."""
|
||||
try:
|
||||
from infrastructure.timmy_bridge.mlx.mlx_integration import MLXInference
|
||||
obj = MLXInference()
|
||||
result = obj.load_model(model_path='/tmp/test')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: infrastructure/timmy-bridge/mlx/mlx_integration.py:93
|
||||
# Function: MLXInference.reflect
|
||||
|
||||
def test_infrastructure_timmy_bridge_mlx_mlx_integration_MLXInference_reflect(self):
|
||||
"""Test MLXInference.reflect -- auto-generated."""
|
||||
try:
|
||||
from infrastructure.timmy_bridge.mlx.mlx_integration import MLXInference
|
||||
obj = MLXInference()
|
||||
result = obj.reflect()
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
|
||||
# AUTO-GENERATED -- review before merging
|
||||
# Source: infrastructure/timmy-bridge/mlx/mlx_integration.py:108
|
||||
# Function: MLXInference.respond_to
|
||||
|
||||
def test_infrastructure_timmy_bridge_mlx_mlx_integration_MLXInference_respond_to(self):
|
||||
"""Test MLXInference.respond_to -- auto-generated."""
|
||||
try:
|
||||
from infrastructure.timmy_bridge.mlx.mlx_integration import MLXInference
|
||||
obj = MLXInference()
|
||||
result = obj.respond_to(message='test msg', context='test msg')
|
||||
assert result is not None or result is None # Placeholder
|
||||
except ImportError:
|
||||
pytest.skip('Module not importable')
|
||||
Reference in New Issue
Block a user