Compare commits
1 Commits
step35/134
...
step35/96-
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
365ab66e88 |
203
scripts/docstring_generator.py
Normal file
203
scripts/docstring_generator.py
Normal file
@@ -0,0 +1,203 @@
|
||||
#!/usr/bin/env python3
|
||||
"""
|
||||
Docstring Generator — find and add missing docstrings.
|
||||
|
||||
Scans Python files for functions/async functions lacking docstrings.
|
||||
Generates Google-style docstrings from function signature and body.
|
||||
Inserts them in place.
|
||||
|
||||
Usage:
|
||||
python3 docstring_generator.py scripts/ # Fix in place
|
||||
python3 docstring_generator.py --dry-run scripts/ # Preview changes
|
||||
python3 docstring_generator.py --json scripts/ # Machine-readable output
|
||||
python3 docstring_generator.py path/to/file.py
|
||||
"""
|
||||
|
||||
import argparse
|
||||
import ast
|
||||
import json
|
||||
import os
|
||||
import sys
|
||||
from pathlib import Path
|
||||
from typing import Optional, Tuple, List
|
||||
|
||||
|
||||
# --- Helper: turn snake_case into Title Case phrase ---
|
||||
def name_to_title(name: str) -> str:
|
||||
"""Convert snake_case function name to a Title Case description."""
|
||||
words = name.replace('_', ' ').split()
|
||||
if not words:
|
||||
return ''
|
||||
titled = []
|
||||
for w in words:
|
||||
if len(w) <= 2:
|
||||
titled.append(w.upper())
|
||||
else:
|
||||
titled.append(w[0].upper() + w[1:])
|
||||
return ' '.join(titled)
|
||||
|
||||
|
||||
# --- Helper: extract first meaningful statement from body for summary ---
|
||||
def extract_body_hint(body: list[ast.stmt]) -> Optional[str]:
|
||||
"""Look for an assignment or return that hints at function purpose."""
|
||||
for stmt in body:
|
||||
if isinstance(stmt, ast.Expr) and isinstance(stmt.value, ast.Constant):
|
||||
continue # skip existing docstring placeholder
|
||||
# Assignment to a result-like variable?
|
||||
if isinstance(stmt, ast.Assign):
|
||||
for target in stmt.targets:
|
||||
if isinstance(target, ast.Name):
|
||||
var_name = target.id
|
||||
if var_name in ('result', 'msg', 'output', 'retval', 'value', 'response', 'data'):
|
||||
val = ast.unparse(stmt.value).strip()
|
||||
if val:
|
||||
return f"Compute or return {val}"
|
||||
# Return statement
|
||||
if isinstance(stmt, ast.Return) and stmt.value:
|
||||
ret = ast.unparse(stmt.value).strip()
|
||||
if ret:
|
||||
return f"Return {ret}"
|
||||
break
|
||||
return None
|
||||
|
||||
|
||||
# --- Generate a docstring string for a function ---
|
||||
def generate_docstring(func_node: ast.FunctionDef | ast.AsyncFunctionDef) -> str:
|
||||
"""Build a Google-style docstring for the given function node."""
|
||||
parts: list[str] = []
|
||||
|
||||
# Summary line
|
||||
summary = name_to_title(func_node.name)
|
||||
body_hint = extract_body_hint(func_node.body)
|
||||
if body_hint:
|
||||
summary = f"{summary}. {body_hint}"
|
||||
parts.append(summary)
|
||||
|
||||
# Args section if there are parameters (excluding self/cls)
|
||||
args = func_node.args.args
|
||||
if args:
|
||||
arg_lines = []
|
||||
for arg in args:
|
||||
if arg.arg in ('self', 'cls'):
|
||||
continue
|
||||
type_ann = ast.unparse(arg.annotation) if arg.annotation else 'Any'
|
||||
arg_lines.append(f"{arg.arg} ({type_ann}): Parameter {arg.arg}")
|
||||
if arg_lines:
|
||||
parts.append("\nArgs:\n " + "\n ".join(arg_lines))
|
||||
|
||||
# Returns section
|
||||
if func_node.returns:
|
||||
ret_type = ast.unparse(func_node.returns)
|
||||
parts.append(f"\nReturns:\n {ret_type}: Return value")
|
||||
elif any(isinstance(s, ast.Return) and s.value is not None for s in ast.walk(func_node)):
|
||||
parts.append("\nReturns:\n Return value")
|
||||
|
||||
return '"""' + '\n'.join(parts) + '\n"""'
|
||||
|
||||
|
||||
# --- Transform source AST ---
|
||||
def process_source(source: str, filename: str) -> Tuple[str, List[str]]:
|
||||
"""Add docstrings to all undocumented functions. Returns (new_source, [func_names])."""
|
||||
try:
|
||||
tree = ast.parse(source)
|
||||
except SyntaxError as e:
|
||||
print(f" WARNING: Could not parse {filename}: {e}", file=sys.stderr)
|
||||
return source, []
|
||||
|
||||
class DocstringInserter(ast.NodeTransformer):
|
||||
def __init__(self):
|
||||
self.modified_funcs: list[str] = []
|
||||
|
||||
def visit_FunctionDef(self, node: ast.FunctionDef) -> ast.FunctionDef:
|
||||
return self._process(node)
|
||||
|
||||
def visit_AsyncFunctionDef(self, node: ast.AsyncFunctionDef) -> ast.AsyncFunctionDef:
|
||||
return self._process(node)
|
||||
|
||||
def _process(self, node):
|
||||
existing_doc = ast.get_docstring(node)
|
||||
if existing_doc is not None:
|
||||
return node
|
||||
docstring_text = generate_docstring(node)
|
||||
doc_node = ast.Expr(value=ast.Constant(value=docstring_text))
|
||||
node.body.insert(0, doc_node)
|
||||
ast.fix_missing_locations(node)
|
||||
self.modified_funcs.append(node.name)
|
||||
return node
|
||||
|
||||
inserter = DocstringInserter()
|
||||
new_tree = inserter.visit(tree)
|
||||
if inserter.modified_funcs:
|
||||
return ast.unparse(new_tree), inserter.modified_funcs
|
||||
return source, []
|
||||
|
||||
|
||||
# --- File discovery ---
|
||||
def iter_python_files(paths: list[str]) -> list[Path]:
|
||||
"""Collect all .py files from provided paths."""
|
||||
files: set[Path] = set()
|
||||
for p in paths:
|
||||
path = Path(p)
|
||||
if not path.exists():
|
||||
print(f"WARNING: Path not found: {p}", file=sys.stderr)
|
||||
continue
|
||||
if path.is_file() and path.suffix == '.py':
|
||||
files.add(path.resolve())
|
||||
elif path.is_dir():
|
||||
for child in path.rglob('*.py'):
|
||||
if '.git' in child.parts or '__pycache__' in child.parts:
|
||||
continue
|
||||
files.add(child.resolve())
|
||||
return sorted(files)
|
||||
|
||||
|
||||
def main():
|
||||
parser = argparse.ArgumentParser(description="Generate docstrings for functions missing them")
|
||||
parser.add_argument('paths', nargs='+', help='Python files or directories to process')
|
||||
parser.add_argument('--dry-run', action='store_true', help='Show what would change without writing')
|
||||
parser.add_argument('--json', action='store_true', help='Output machine-readable JSON summary')
|
||||
parser.add_argument('-v', '--verbose', action='store_true', help='Print each file processed')
|
||||
|
||||
args = parser.parse_args()
|
||||
|
||||
files = iter_python_files(args.paths)
|
||||
if not files:
|
||||
print("No Python files found to process", file=sys.stderr)
|
||||
sys.exit(1)
|
||||
|
||||
results = []
|
||||
total_funcs = 0
|
||||
|
||||
for pyfile in files:
|
||||
try:
|
||||
original = pyfile.read_text(encoding='utf-8')
|
||||
except Exception as e:
|
||||
print(f" ERROR reading {pyfile}: {e}", file=sys.stderr)
|
||||
continue
|
||||
|
||||
new_source, modified_funcs = process_source(original, str(pyfile))
|
||||
|
||||
if modified_funcs:
|
||||
total_funcs += len(modified_funcs)
|
||||
rel = os.path.relpath(pyfile)
|
||||
if args.verbose:
|
||||
print(f" {rel}: +{len(modified_funcs)} docstrings")
|
||||
results.append({'file': str(pyfile), 'functions': modified_funcs})
|
||||
if not args.dry_run:
|
||||
pyfile.write_text(new_source, encoding='utf-8')
|
||||
elif args.verbose:
|
||||
print(f" {rel}: no changes")
|
||||
|
||||
if args.json:
|
||||
summary = {'total_files_modified': len(results), 'total_functions': total_funcs, 'files': results}
|
||||
print(json.dumps(summary, indent=2))
|
||||
else:
|
||||
print(f"Generated docstrings for {total_funcs} functions across {len(results)} files")
|
||||
if args.dry_run:
|
||||
print(" (dry run — no files written)")
|
||||
|
||||
return 0
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
sys.exit(main())
|
||||
@@ -1,258 +0,0 @@
|
||||
#!/usr/bin/env python3
|
||||
"""GitHub Trending Scanner — Scan trending repos in AI/ML.
|
||||
|
||||
Extracts: repo description, stars, key features (topics, inferred highlights).
|
||||
Filters by language and/or topic. Outputs dated JSON for daily scan pipeline.
|
||||
|
||||
Usage:
|
||||
python3 github_trending_scanner.py --language python --topic ai --output metrics/trending
|
||||
python3 github_trending_scanner.py --topic machine-learning --limit 50
|
||||
python3 github_trending_scanner.py --language rust --topic artificial-intelligence
|
||||
"""
|
||||
|
||||
import argparse
|
||||
import json
|
||||
import os
|
||||
import sys
|
||||
import time
|
||||
from datetime import datetime, timezone
|
||||
from pathlib import Path
|
||||
from typing import Optional, List, Dict
|
||||
import urllib.request
|
||||
import urllib.parse
|
||||
import urllib.error
|
||||
|
||||
GITHUB_API_BASE = os.environ.get("GITHUB_API_BASE", "https://api.github.com")
|
||||
DEFAULT_OUTPUT_DIR = os.environ.get("TRENDING_OUTPUT_DIR", "metrics/trending")
|
||||
DEFAULT_LIMIT = int(os.environ.get("TRENDING_LIMIT", "30"))
|
||||
DEFAULT_MIN_STARS = int(os.environ.get("TRENDING_MIN_STARS", "1000"))
|
||||
|
||||
|
||||
def fetch_trending_repos(
|
||||
language: Optional[str] = None,
|
||||
topic: Optional[str] = None,
|
||||
min_stars: int = DEFAULT_MIN_STARS,
|
||||
limit: int = DEFAULT_LIMIT,
|
||||
) -> List[Dict]:
|
||||
"""Fetch trending-like repositories from GitHub using the search API.
|
||||
|
||||
GitHub's public search API is unauthenticated-rate-limited (60 req/hr).
|
||||
This function retries on rate-limit backoff and falls back gracefully.
|
||||
"""
|
||||
# Build search query: stars threshold + optional language/topic filters
|
||||
query = f"stars:>{min_stars}"
|
||||
if language:
|
||||
query += f" language:{language}"
|
||||
if topic:
|
||||
query += f" topic:{topic}"
|
||||
|
||||
# Sort by stars descending as a proxy for trending/popular
|
||||
params = {
|
||||
"q": query,
|
||||
"sort": "stars",
|
||||
"order": "desc",
|
||||
"per_page": min(limit, 100), # GitHub max per_page is 100
|
||||
}
|
||||
url = f"{GITHUB_API_BASE}/search/repositories?{urllib.parse.urlencode(params)}"
|
||||
|
||||
headers = {
|
||||
"Accept": "application/vnd.github.v3+json",
|
||||
"User-Agent": "Sovereign-Trending-Scanner/1.0",
|
||||
}
|
||||
|
||||
for attempt in range(3):
|
||||
try:
|
||||
req = urllib.request.Request(url, headers=headers)
|
||||
with urllib.request.urlopen(req, timeout=30) as resp:
|
||||
if resp.status != 200:
|
||||
raise RuntimeError(f"GitHub API returned {resp.status}")
|
||||
data = json.loads(resp.read().decode("utf-8"))
|
||||
return data.get("items", [])[:limit]
|
||||
except urllib.error.HTTPError as e:
|
||||
if e.code == 403:
|
||||
# Check for rate limit message
|
||||
body = e.read().decode("utf-8", errors="replace").lower()
|
||||
if "rate limit" in body or "api rate limit exceeded" in body:
|
||||
reset_ts = int(e.headers.get("X-RateLimit-Reset", 0))
|
||||
wait_seconds = max(5, reset_ts - int(time.time()) + 5)
|
||||
print(f"Rate limit exceeded — waiting {wait_seconds}s (attempt {attempt+1}/3)...", file=sys.stderr)
|
||||
time.sleep(wait_seconds)
|
||||
continue
|
||||
print(f"ERROR: GitHub API request failed: {e} — {e.read().decode('utf-8', errors='replace')[:200]}", file=sys.stderr)
|
||||
return []
|
||||
except Exception as e:
|
||||
if attempt < 2:
|
||||
backoff = 2 ** attempt
|
||||
print(f"WARNING: Fetch attempt {attempt+1} failed: {e} — retrying in {backoff}s", file=sys.stderr)
|
||||
time.sleep(backoff)
|
||||
continue
|
||||
print(f"ERROR: All fetch attempts failed: {e}", file=sys.stderr)
|
||||
return []
|
||||
|
||||
return []
|
||||
|
||||
|
||||
def extract_repo_features(repo_data: Dict) -> Dict:
|
||||
"""Extract structured fields for a trending repo."""
|
||||
description = (repo_data.get("description") or "").strip()
|
||||
topics = repo_data.get("topics", [])
|
||||
|
||||
# Infer key features from description and topics
|
||||
features = infer_features(description, topics)
|
||||
|
||||
return {
|
||||
"name": repo_data.get("full_name", ""),
|
||||
"description": description,
|
||||
"stars": repo_data.get("stargazers_count", 0),
|
||||
"forks": repo_data.get("forks_count", 0),
|
||||
"open_issues": repo_data.get("open_issues_count", 0),
|
||||
"language": repo_data.get("language", ""),
|
||||
"topics": topics,
|
||||
"url": repo_data.get("html_url", ""),
|
||||
"created_at": repo_data.get("created_at", ""),
|
||||
"updated_at": repo_data.get("updated_at", ""),
|
||||
"key_features": features,
|
||||
"scanned_at": datetime.now(timezone.utc).isoformat(),
|
||||
}
|
||||
|
||||
|
||||
def infer_features(description: str, topics: List[str]) -> List[str]:
|
||||
"""Infer notable capabilities/features from repo metadata.
|
||||
|
||||
Looks for AI/ML-relevant capabilities in topics and description.
|
||||
"""
|
||||
features = []
|
||||
text = (description + " " + " ".join(topics)).lower()
|
||||
|
||||
# Domain capabilities (keys normalized to lowercase for consistency)
|
||||
capability_keywords = {
|
||||
"fine-tuning": ["fine-tun", "finetun"],
|
||||
"agent framework": ["agent"],
|
||||
"local/offline": ["local", "on-device", "offline"],
|
||||
"quantized models": ["quantized", "quantization", "gguf", "gptq"],
|
||||
"vision": ["vision", "multimodal", "image", "visual"],
|
||||
"speech/audio": ["speech", "audio", "whisper", "tts"],
|
||||
"retrieval/rag": ["rag", "retrieval", "embedding", "vector"],
|
||||
"training": ["train", "training", "sft", "dpo"],
|
||||
"gui/playground": ["gui", "playground", "webui", "interface"],
|
||||
"sota": ["state-of-the-art", "sota", "latest"],
|
||||
}
|
||||
|
||||
for label, keywords in capability_keywords.items():
|
||||
if any(kw in text for kw in keywords):
|
||||
features.append(label)
|
||||
|
||||
# Also include non-generic topics as features
|
||||
generic_topics = {"ai", "ml", "machine-learning", "deep-learning", "llm", "python", "pytorch", "tensorflow"}
|
||||
for topic in topics:
|
||||
if topic.lower() not in generic_topics:
|
||||
features.append(topic)
|
||||
|
||||
# Deduplicate while preserving order, return up to 10
|
||||
seen = set()
|
||||
unique = []
|
||||
for f in features:
|
||||
key = f.lower()
|
||||
if key not in seen:
|
||||
seen.add(key)
|
||||
unique.append(f)
|
||||
return unique[:10]
|
||||
|
||||
|
||||
def save_trending(repos: List[Dict], output_dir: str = "metrics/trending") -> str:
|
||||
"""Save trending results to a dated JSON file.
|
||||
|
||||
Returns the path of the written file.
|
||||
"""
|
||||
output_path = Path(output_dir)
|
||||
output_path.mkdir(parents=True, exist_ok=True)
|
||||
|
||||
date_str = datetime.now(timezone.utc).strftime("%Y-%m-%d")
|
||||
filename = output_path / f"github-trending-{date_str}.json"
|
||||
|
||||
output_data = {
|
||||
"scanned_at": datetime.now(timezone.utc).isoformat(),
|
||||
"count": len(repos),
|
||||
"repos": repos,
|
||||
}
|
||||
|
||||
with open(filename, "w") as f:
|
||||
json.dump(output_data, f, indent=2, ensure_ascii=False)
|
||||
|
||||
return str(filename)
|
||||
|
||||
|
||||
def main() -> None:
|
||||
parser = argparse.ArgumentParser(
|
||||
description="Scan GitHub trending repositories in AI/ML"
|
||||
)
|
||||
parser.add_argument(
|
||||
"--language",
|
||||
help="Filter by programming language (e.g., python, rust, go)",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--topic",
|
||||
help="Filter by GitHub topic (e.g., ai, machine-learning, llm)",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--since",
|
||||
default="daily",
|
||||
choices=["daily", "weekly", "monthly"],
|
||||
help="Trending period (daily/weekly/monthly) — informational only",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--output",
|
||||
default="metrics/trending",
|
||||
help="Output directory for results (default: metrics/trending)",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--limit",
|
||||
type=int,
|
||||
default=DEFAULT_LIMIT,
|
||||
help=f"Maximum repos to fetch (default: {DEFAULT_LIMIT})",
|
||||
)
|
||||
parser.add_argument(
|
||||
"--min-stars",
|
||||
type=int,
|
||||
default=DEFAULT_MIN_STARS,
|
||||
help=f"Minimum star count for relevance (default: {DEFAULT_MIN_STARS})",
|
||||
)
|
||||
args = parser.parse_args()
|
||||
|
||||
print(
|
||||
f"Fetching trending repos "
|
||||
f"(language={args.language or 'any'}, topic={args.topic or 'any'}, period={args.since})..."
|
||||
)
|
||||
|
||||
repos_raw = fetch_trending_repos(
|
||||
language=args.language,
|
||||
topic=args.topic,
|
||||
min_stars=args.min_stars,
|
||||
limit=args.limit,
|
||||
)
|
||||
|
||||
if not repos_raw:
|
||||
print("WARNING: No repos fetched — check network or rate limits", file=sys.stderr)
|
||||
|
||||
repos = [extract_repo_features(r) for r in repos_raw]
|
||||
|
||||
output_file = save_trending(repos, args.output)
|
||||
print(f"Saved {len(repos)} trending repos to {output_file}")
|
||||
|
||||
# Brief human-readable summary
|
||||
if repos:
|
||||
print("\nTop repos:")
|
||||
for repo in repos[:5]:
|
||||
features_preview = ", ".join(repo["key_features"][:3])
|
||||
print(f" ★ {repo['stars']:>7} {repo['name']}")
|
||||
if repo["description"]:
|
||||
desc = repo["description"][:80]
|
||||
print(f" {desc}{'...' if len(repo['description']) > 80 else ''}")
|
||||
if features_preview:
|
||||
print(f" Features: {features_preview}")
|
||||
|
||||
return 0
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
sys.exit(main())
|
||||
@@ -1,125 +0,0 @@
|
||||
#!/usr/bin/env python3
|
||||
"""Tests for github_trending_scanner.py — pure function validation.
|
||||
|
||||
Tests the feature inference, extraction, and output formatting logic
|
||||
without relying on external GitHub API calls.
|
||||
"""
|
||||
|
||||
import json
|
||||
import sys
|
||||
import tempfile
|
||||
from pathlib import Path
|
||||
|
||||
# Add scripts dir to path for import
|
||||
sys.path.insert(0, str(Path(__file__).resolve().parent))
|
||||
|
||||
from github_trending_scanner import (
|
||||
extract_repo_features,
|
||||
infer_features,
|
||||
save_trending,
|
||||
)
|
||||
|
||||
|
||||
def test_infer_features_from_description():
|
||||
"""Feature inference extracts capabilities from description text."""
|
||||
desc = "A local, quantized LLM framework for fine-tuning and agent-based RAG with vision."
|
||||
topics = ["ai", "llm"]
|
||||
features = infer_features(desc, topics)
|
||||
|
||||
# Should include relevant capabilities (case-insensitive comparison)
|
||||
expected_lower = {"fine-tuning", "local/offline", "quantized models", "agent framework", "vision", "retrieval/rag"}
|
||||
actual_lower = set(f.lower() for f in features)
|
||||
assert expected_lower.issubset(actual_lower), f"Missing features. Expected subset of {expected_lower}, got {actual_lower}"
|
||||
print("PASS: infer_features_from_description")
|
||||
|
||||
|
||||
def test_infer_features_from_topics_only():
|
||||
"""Topics alone can drive feature detection."""
|
||||
desc = ""
|
||||
topics = ["computer-vision", "speech", "pytorch"]
|
||||
features = infer_features(desc, topics)
|
||||
|
||||
# Non-generic topics should appear as features (topics preserved as-is)
|
||||
assert "computer-vision" in features, f"Expected 'computer-vision' in {features}"
|
||||
assert "speech" in features, f"Expected 'speech' in {features}"
|
||||
# Generic topics (pytorch) may be filtered
|
||||
print(f"PASS: infer_features_from_topics_only → {features}")
|
||||
|
||||
|
||||
def test_extract_repo_features_produces_valid_structure():
|
||||
"""extract_repo_features returns all required fields."""
|
||||
mock_repo = {
|
||||
"full_name": "example/repo",
|
||||
"description": "An example repository",
|
||||
"stargazers_count": 1234,
|
||||
"forks_count": 56,
|
||||
"open_issues_count": 7,
|
||||
"language": "Python",
|
||||
"topics": ["ai", "llm"],
|
||||
"html_url": "https://github.com/example/repo",
|
||||
"created_at": "2025-01-01T00:00:00Z",
|
||||
"updated_at": "2026-01-01T00:00:00Z",
|
||||
}
|
||||
|
||||
result = extract_repo_features(mock_repo)
|
||||
|
||||
assert result["name"] == "example/repo"
|
||||
assert result["description"] == "An example repository"
|
||||
assert result["stars"] == 1234
|
||||
assert isinstance(result["key_features"], list)
|
||||
assert "scanned_at" in result
|
||||
assert result["url"] == "https://github.com/example/repo"
|
||||
print("PASS: extract_repo_features_structure")
|
||||
|
||||
|
||||
def test_save_trending_creates_dated_json():
|
||||
"""save_trending writes a valid JSON file with the expected schema."""
|
||||
repos = [
|
||||
{
|
||||
"name": "test/repo",
|
||||
"description": "Test repository",
|
||||
"stars": 999,
|
||||
"language": "Python",
|
||||
"topics": ["test"],
|
||||
"key_features": ["testing"],
|
||||
"scanned_at": "2026-04-26T00:00:00+00:00",
|
||||
}
|
||||
]
|
||||
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
output_file = save_trending(repos, output_dir=tmp)
|
||||
|
||||
path = Path(output_file)
|
||||
assert path.exists(), f"Output file not created: {output_file}"
|
||||
|
||||
with open(path) as f:
|
||||
data = json.load(f)
|
||||
|
||||
assert "scanned_at" in data
|
||||
assert data["count"] == 1
|
||||
assert isinstance(data["repos"], list)
|
||||
assert data["repos"][0]["name"] == "test/repo"
|
||||
print(f"PASS: save_trending → {output_file}")
|
||||
|
||||
|
||||
def test_save_trending_respects_output_dir_creation():
|
||||
"""Output directory is created if it doesn't exist."""
|
||||
repos = []
|
||||
|
||||
with tempfile.TemporaryDirectory() as tmp:
|
||||
nested = Path(tmp) / "nested" / "trending"
|
||||
assert not nested.exists()
|
||||
|
||||
output_file = save_trending(repos, output_dir=str(nested))
|
||||
assert nested.exists()
|
||||
assert Path(output_file).exists()
|
||||
print("PASS: output_dir_creation")
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
test_infer_features_from_description()
|
||||
test_infer_features_from_topics_only()
|
||||
test_extract_repo_features_produces_valid_structure()
|
||||
test_save_trending_creates_dated_json()
|
||||
test_save_trending_respects_output_dir_creation()
|
||||
print("\nAll github_trending_scanner tests passed.")
|
||||
128
tests/test_docstring_generator.py
Normal file
128
tests/test_docstring_generator.py
Normal file
@@ -0,0 +1,128 @@
|
||||
"""Tests for docstring_generator module (Issue #96)."""
|
||||
|
||||
import ast
|
||||
import sys
|
||||
import tempfile
|
||||
from pathlib import Path
|
||||
|
||||
import pytest
|
||||
|
||||
sys.path.insert(0, str(Path(__file__).parent.parent / "scripts"))
|
||||
|
||||
from docstring_generator import (
|
||||
name_to_title,
|
||||
extract_body_hint,
|
||||
generate_docstring,
|
||||
process_source,
|
||||
iter_python_files,
|
||||
)
|
||||
|
||||
|
||||
class TestNameToTitle:
|
||||
def test_snake_to_title(self):
|
||||
assert name_to_title("validate_fact") == "Validate Fact"
|
||||
assert name_to_title("docstring_generator") == "Docstring Generator"
|
||||
assert name_to_title("main") == "Main"
|
||||
assert name_to_title("__init__") == "Init"
|
||||
|
||||
|
||||
class TestExtractBodyHint:
|
||||
def test_assignment_hint(self):
|
||||
body = [ast.parse("result = compute()").body[0]]
|
||||
hint = extract_body_hint(body)
|
||||
assert hint == "Compute or return compute()"
|
||||
|
||||
def test_return_hint(self):
|
||||
body = [ast.parse("return data").body[0]]
|
||||
hint = extract_body_hint(body)
|
||||
assert hint == "Return data"
|
||||
|
||||
def test_no_hint(self):
|
||||
body = [ast.parse("pass").body[0]]
|
||||
assert extract_body_hint(body) is None
|
||||
|
||||
|
||||
class TestGenerateDocstring:
|
||||
def test_simple_function(self):
|
||||
src = "def add(a, b):\n return a + b\n"
|
||||
tree = ast.parse(src)
|
||||
func = tree.body[0]
|
||||
doc = generate_docstring(func)
|
||||
assert 'Add' in doc
|
||||
assert 'a' in doc and 'b' in doc
|
||||
assert 'Args:' in doc
|
||||
assert 'Returns:' in doc
|
||||
|
||||
def test_typed_function(self):
|
||||
src = "def greet(name: str) -> str:\n return f'Hello {name}'\n"
|
||||
tree = ast.parse(src)
|
||||
func = tree.body[0]
|
||||
doc = generate_docstring(func)
|
||||
assert 'name (str)' in doc
|
||||
assert 'str' in doc
|
||||
|
||||
def test_async_function(self):
|
||||
src = "async def fetch():\n pass\n"
|
||||
tree = ast.parse(src)
|
||||
func = tree.body[0]
|
||||
doc = generate_docstring(func)
|
||||
assert 'Fetch' in doc
|
||||
|
||||
def test_self_skipped(self):
|
||||
src = "class C:\n def method(self, x):\n return x\n"
|
||||
tree = ast.parse(src)
|
||||
cls = tree.body[0]
|
||||
method = cls.body[0]
|
||||
doc = generate_docstring(method)
|
||||
# 'self' should not appear in Args section
|
||||
args_start = doc.find('Args:')
|
||||
if args_start >= 0:
|
||||
args_section = doc[args_start:]
|
||||
assert '(self)' not in args_section
|
||||
|
||||
|
||||
class TestProcessSource:
|
||||
def test_adds_docstrings(self):
|
||||
src = "def foo(x):\n return x * 2\n"
|
||||
new_src, funcs = process_source(src, "test.py")
|
||||
assert len(funcs) == 1 and funcs[0] == "foo"
|
||||
assert '"""' in new_src
|
||||
assert 'Foo' in new_src
|
||||
|
||||
def test_preserves_existing_docstrings(self):
|
||||
src = 'def bar():\n """Already documented."""\n return 1\n'
|
||||
new_src, funcs = process_source(src, "test.py")
|
||||
assert len(funcs) == 0
|
||||
assert new_src == src
|
||||
|
||||
def test_multiple_functions(self):
|
||||
src = "def a(): pass\ndef b(): pass\ndef c(): pass\n"
|
||||
new_src, funcs = process_source(src, "test.py")
|
||||
assert len(funcs) == 3
|
||||
assert '"""' in new_src
|
||||
|
||||
def test_dry_run_no_write(self, tmp_path):
|
||||
file = tmp_path / "t.py"
|
||||
file.write_text("def f(): pass\n")
|
||||
original_mtime = file.stat().st_mtime
|
||||
new_src, funcs = process_source(file.read_text(), str(file))
|
||||
assert funcs # detected
|
||||
# When caller handles write, dry-run leaves file unchanged
|
||||
current_mtime = file.stat().st_mtime
|
||||
assert current_mtime == original_mtime
|
||||
|
||||
|
||||
class TestIterPythonFiles:
|
||||
def test_single_file(self, tmp_path):
|
||||
f = tmp_path / "single.py"
|
||||
f.write_text("x = 1")
|
||||
files = iter_python_files([str(f)])
|
||||
assert len(files) == 1
|
||||
assert files[0].name == "single.py"
|
||||
|
||||
def test_directory_recursion(self, tmp_path):
|
||||
(tmp_path / "sub").mkdir()
|
||||
(tmp_path / "sub" / "a.py").write_text("a=1")
|
||||
(tmp_path / "b.py").write_text("b=2")
|
||||
files = iter_python_files([str(tmp_path)])
|
||||
assert len(files) == 2
|
||||
Reference in New Issue
Block a user