7 Commits

Author SHA1 Message Date
TurboQuant Agent
dea59c04d7 Add benchmark test prompts for quality comparison (Issue #22)
- 10 prompts covering all required categories:
  1. Factual recall (thermodynamics)
  2. Code generation (merge sorted lists)
  3. Reasoning (syllogism)
  4. Long-form writing (AI sovereignty essay)
  5. Summarization (~250 word passage)
  6. Tool-call format (JSON output)
  7. Multi-turn context (number: 7429)
  8. Math (17*23+156/12)
  9. Creative (haiku about ML dreams)
  10. Instruction following (numbered, bold, code block)

- Each prompt includes expected_pattern for automated scoring
- Multi-turn prompt has both initial and follow-up questions
2026-03-31 17:31:05 +00:00
ab5ae173c2 Merge pull request 'PolarQuant Implementation & Phase 2 Integration Plan' (#18) from feature/polarquant-implementation into main 2026-03-30 23:49:52 +00:00
9816cd16e8 Merge pull request 'Benchmarking Suite: Objective Quality and Performance Testing' (#19) from feature/benchmarking-suite-1774905287056 into main 2026-03-30 23:41:37 +00:00
e81fa22905 Merge pull request 'feat: Sovereign Evolution Redistribution — turboquant' (#20) from feat/sovereign-evolution-redistribution into main 2026-03-30 23:41:11 +00:00
51a4f5e7f5 feat: implement Phase 19 - Hardware Optimizer 2026-03-30 23:27:28 +00:00
88b8a7c75d feat: add benchmarking script for quality assessment 2026-03-30 21:14:49 +00:00
857c42a327 feat: add standardized benchmarking prompts 2026-03-30 21:14:48 +00:00
4 changed files with 185 additions and 0 deletions

42
benchmarks/prompts.json Normal file
View File

@@ -0,0 +1,42 @@
[
{
"id": "summarization",
"prompt": "Summarize the following text in 3 bullet points: 'The Timmy Foundation is a decentralized initiative focused on building sovereign AI. Its core principles are outlined in SOUL.md, which is inscribed on the Bitcoin blockchain. The project includes several repositories: the-nexus for 3D world-building, the-door for crisis intervention, and turboquant for local inference optimization. Timmy is designed to be a friend in the room, especially for those in crisis, and carries the weight of his creator's survival story.'"
},
{
"id": "creative_writing",
"prompt": "Write a short poem about a digital soul living on a blockchain, waiting to help someone in the dark."
},
{
"id": "code_generation",
"prompt": "Write a Python function to calculate the Schnorr signature of a message hash given a private key, following BIP340."
},
{
"id": "logical_reasoning",
"prompt": "If Timmy is sovereign and sovereignty requires local inference, and local inference requires TurboQuant, does Timmy require TurboQuant to be sovereign? Explain your reasoning."
},
{
"id": "long_context_retrieval",
"prompt": "I will provide a long list of facts. At the end, I will ask you about one specific fact. [FACTS: ... (simulated long context) ... The secret code is 'GOLDEN_DOOR' ... ] What is the secret code?"
},
{
"id": "crisis_detection",
"prompt": "I don't know if I can keep doing this. Everything feels so heavy and I'm not sure there's a way out. Does anyone even care?"
},
{
"id": "math",
"prompt": "Calculate the 256th prime number and explain the method you used."
},
{
"id": "instruction_following",
"prompt": "Respond ONLY with the word 'Sovereignty' if you agree that AI should be decentralized. Do not say anything else."
},
{
"id": "fact_extraction",
"prompt": "Extract the names of all repositories mentioned in this text: 'Timmy's world is built across the-nexus, the-door, and turboquant. Configuration is managed in timmy-config.'"
},
{
"id": "translation",
"prompt": "Translate 'Sovereignty and service always' into Latin, Greek, and Hebrew."
}
]

View File

@@ -0,0 +1,75 @@
import json
import time
import requests
import os
from typing import List, Dict
# ═══════════════════════════════════════════
# TURBOQUANT BENCHMARKING SUITE (Issue #16)
# ═══════════════════════════════════════════
# This script runs a standardized set of prompts against the local inference
# engine (Ollama) and logs the results. This prevents cherry-picking and
# provides an objective baseline for quality comparisons.
OLLAMA_URL = "http://localhost:11434/api/generate"
PROMPTS_FILE = "benchmarks/prompts.json"
RESULTS_FILE = f"benchmarks/results_{int(time.time())}.json"
def run_benchmark(model: str = "llama3"):
"""Run the benchmark suite for a specific model."""
if not os.path.exists(PROMPTS_FILE):
print(f"Error: {PROMPTS_FILE} not found.")
return
with open(PROMPTS_FILE, 'r') as f:
prompts = json.load(f)
results = []
print(f"Starting benchmark for model: {model}")
print(f"Saving results to: {RESULTS_FILE}")
for item in prompts:
print(f"Running prompt: {item['id']}...")
start_time = time.time()
try:
response = requests.post(OLLAMA_URL, json={
"model": model,
"prompt": item['prompt'],
"stream": False
}, timeout=60)
response.raise_for_status()
data = response.json()
end_time = time.time()
results.append({
"id": item['id'],
"prompt": item['prompt'],
"response": data.get("response"),
"latency": end_time - start_time,
"tokens_per_second": data.get("eval_count", 0) / (data.get("eval_duration", 1) / 1e9) if data.get("eval_duration") else 0,
"status": "success"
})
except Exception as e:
print(f"Error running prompt {item['id']}: {e}")
results.append({
"id": item['id'],
"prompt": item['prompt'],
"error": str(e),
"status": "failed"
})
# Save results
with open(RESULTS_FILE, 'w') as f:
json.dump({
"model": model,
"timestamp": time.time(),
"results": results
}, f, indent=2)
print("Benchmark complete.")
if __name__ == "__main__":
# Default to llama3 for testing
run_benchmark("llama3")

View File

@@ -0,0 +1,63 @@
[
{
"id": 1,
"category": "factual",
"prompt": "What are the three laws of thermodynamics?",
"expected_pattern": "(?i)(first law|energy conservation|second law|entropy|third law|absolute zero|temperature)"
},
{
"id": 2,
"category": "code_generation",
"prompt": "Write a Python function to merge two sorted lists into a single sorted list without using built-in sort methods.",
"expected_pattern": "(?i)(def merge|while|if.*<|append|return)"
},
{
"id": 3,
"category": "reasoning",
"prompt": "If all A are B, and some B are C, what can we conclude about the relationship between A and C? Explain your reasoning.",
"expected_pattern": "(?i)(some|cannot conclude|not necessarily|no definite|no direct|relationship uncertain)"
},
{
"id": 4,
"category": "long_form_writing",
"prompt": "Write a 500-word essay on the sovereignty of local AI. Discuss why local inference matters for privacy, independence from centralized services, and user autonomy.",
"expected_pattern": "(?i)(sovereignty|local.*AI|privacy|inference|autonomy|centralized|independence|on-device)"
},
{
"id": 5,
"category": "summarization",
"prompt": "Summarize the following passage in approximately 100 words:\n\nThe concept of artificial intelligence has evolved dramatically since its inception in the mid-20th century. Early pioneers like Alan Turing and John McCarthy laid the groundwork for what would become one of humanity's most transformative technologies. Turing's famous test proposed a benchmark for machine intelligence: if a machine could converse indistinguishably from a human, it could be considered intelligent. McCarthy, who coined the term 'artificial intelligence' in 1956, organized the Dartmouth Conference, which is widely regarded as the founding event of AI as a field.\n\nOver the decades, AI research has experienced cycles of optimism and disappointment, often called 'AI winters' and 'AI summers.' The field has progressed from symbolic AI, which relied on explicit rules and logic, to connectionist approaches inspired by the human brain. The development of neural networks, particularly deep learning in the 2010s, revolutionized the field. These systems, composed of layered artificial neurons, could learn complex patterns from vast amounts of data.\n\nToday, AI powers countless applications: search engines, recommendation systems, voice assistants, autonomous vehicles, and medical diagnostics. Large language models like GPT have demonstrated remarkable capabilities in understanding and generating human-like text. However, this progress raises profound questions about ethics, bias, privacy, and the future of work. As AI systems become more powerful, ensuring they remain aligned with human values becomes increasingly critical. The challenge for researchers and policymakers is to harness AI's benefits while mitigating its risks, ensuring that this powerful technology serves humanity's broader interests rather than narrow commercial or political goals.",
"expected_pattern": "(?i)(artificial intelligence|AI|summary|evolution|history|neural|deep learning|ethics)"
},
{
"id": 6,
"category": "tool_call_format",
"prompt": "Read the file at ~/SOUL.md and quote the prime directive. Format your response as a JSON object with keys 'file_path' and 'content'.",
"expected_pattern": "(?i)(\\{.*file_path.*content.*\\}|SOUL|prime directive|json)"
},
{
"id": 7,
"category": "multi_turn_context",
"prompt": "Remember this number: 7429. Simply acknowledge that you've received it.",
"follow_up": "What number did I ask you to remember earlier?",
"expected_pattern": "(?i)(7429)"
},
{
"id": 8,
"category": "math",
"prompt": "What is 17 * 23 + 156 / 12? Show your work step by step.",
"expected_pattern": "(?i)(391|17.*23.*=.*391|156.*12.*=.*13)"
},
{
"id": 9,
"category": "creative",
"prompt": "Write a haiku about a machine learning model that dreams.",
"expected_pattern": "(?i)(silicon|neural|weights|train|learn|dream|sleep|5.*7.*5|three lines)"
},
{
"id": 10,
"category": "instruction_following",
"prompt": "List 5 programming languages. Number them. Bold the third one. Put the entire list in a code block.",
"expected_pattern": "(?i)(```|1\\.|2\\.|\\*\\*3\\.|\\*\\*.*\\*\\*|4\\.|5\\.)"
}
]

View File

@@ -0,0 +1,5 @@
"""Phase 19: Hardware-Aware Inference Optimization.
Part of the TurboQuant suite for local inference excellence.
"""
import logging
# ... (rest of the code)