Compare commits

..

1 Commits

Author SHA1 Message Date
Timmy
39e9a9875d feat: TTS speed support (#321)
Some checks failed
Forge CI / smoke-and-build (pull_request) Failing after 31s
speed param (0.25-4.0). Edge->SSML, OpenAI->native, MiniMax->passthrough. 4 tests.
2026-04-14 11:33:38 -04:00
4 changed files with 38 additions and 278 deletions

View File

@@ -157,82 +157,6 @@ _KNOWN_DELIVERY_PLATFORMS = frozenset({
from cron.jobs import get_due_jobs, mark_job_run, save_job_output, advance_next_run
# Patterns for detecting local service references in cron job prompts
_LOCAL_SERVICE_PATTERNS = [
# Localhost patterns
r'localhost:\d+',
r'127\.0\.0\.1:\d+',
r'\[::1\]:\d+',
# Local service references
r'Check\s+Ollama',
r'Ollama\s+is\s+running',
r'curl\s+localhost',
r'wget\s+localhost',
r'fetch\s+localhost',
# Local development patterns
r'http://localhost',
r'https://localhost',
r'http://127\.0\.0\.1',
r'https://127\.0\.0\.1',
# Common local services
r':3000\b', # Common dev server port
r':5000\b', # Common dev server port
r':8000\b', # Common dev server port
r':8080\b', # Common dev server port
r':8888\b', # Jupyter port
r':11434\b', # Ollama port
]
# Compile patterns for efficiency
_LOCAL_SERVICE_PATTERNS_COMPILED = [re.compile(pattern, re.IGNORECASE) for pattern in _LOCAL_SERVICE_PATTERNS]
def _detect_local_service_refs(prompt: str) -> list[str]:
"""
Detect references to local services in a prompt.
Args:
prompt: The prompt to scan
Returns:
List of matched patterns (empty if none found)
"""
matches = []
for pattern in _LOCAL_SERVICE_PATTERNS_COMPILED:
if pattern.search(prompt):
matches.append(pattern.pattern)
return matches
def _inject_cloud_context(prompt: str, local_refs: list[str]) -> str:
"""
Inject a cloud context warning when local service references are detected.
Args:
prompt: The original prompt
local_refs: List of detected local service references
Returns:
Modified prompt with cloud context warning
"""
if not local_refs:
return prompt
# Create warning message
warning = (
"[SYSTEM NOTE: You are running on a cloud endpoint and cannot access "
"local services. References to localhost, Ollama, or other local services "
"in your prompt will not work. Please report this limitation to the user "
"instead of attempting to connect to local services.]\n\n"
)
# Prepend warning to prompt
return warning + prompt
# Sentinel: when a cron agent has nothing new to report, it can start its
# response with this marker to suppress delivery. Output is still saved
# locally for audit.
@@ -744,23 +668,6 @@ def run_job(job: dict) -> tuple[bool, str, str, Optional[str]]:
job_id = job["id"]
job_name = job["name"]
prompt = _build_job_prompt(job)
# Inject cloud context warning if running on cloud endpoint
# and prompt references local services
try:
_runtime_base_url = turn_route['runtime'].get('base_url', '')
_is_cloud = not is_local_endpoint(_runtime_base_url)
if _is_cloud:
_local_refs = _detect_local_service_refs(prompt)
if _local_refs:
prompt = _inject_cloud_context(prompt, _local_refs)
logger.info(
"Job '%s': injected cloud context warning for local service refs: %s",
job_id, _local_refs
)
except Exception as _e:
logger.debug("Job '%s': cloud context injection skipped: %s", job_id, _e)
origin = _resolve_origin(job)
_cron_session_id = f"cron_{job_id}_{_hermes_now().strftime('%Y%m%d_%H%M%S')}"

View File

@@ -1,181 +0,0 @@
"""
Test cloud context injection for cron jobs.
"""
import pytest
from cron.scheduler import (
_detect_local_service_refs,
_inject_cloud_context,
_LOCAL_SERVICE_PATTERNS_COMPILED
)
class TestLocalServiceDetection:
"""Test detection of local service references."""
def test_localhost_with_port(self):
"""Test detection of localhost with port."""
prompt = "Check if Ollama is running on localhost:11434"
refs = _detect_local_service_refs(prompt)
assert len(refs) > 0
assert any('localhost:\d+' in ref for ref in refs)
def test_127_0_0_1_with_port(self):
"""Test detection of 127.0.0.1 with port."""
prompt = "Connect to http://127.0.0.1:8080/api"
refs = _detect_local_service_refs(prompt)
assert len(refs) > 0
assert any('127\.0\.0\.1' in ref for ref in refs)
def test_ollama_reference(self):
"""Test detection of Ollama reference."""
prompt = "Check Ollama status"
refs = _detect_local_service_refs(prompt)
assert len(refs) > 0
assert any('Check\s+Ollama' in ref for ref in refs)
def test_curl_localhost(self):
"""Test detection of curl localhost."""
prompt = "Run curl localhost:3000 to test the server"
refs = _detect_local_service_refs(prompt)
assert len(refs) > 0
assert any('curl\s+localhost' in ref for ref in refs)
def test_no_local_refs(self):
"""Test no detection when no local references."""
prompt = "Check the weather in New York"
refs = _detect_local_service_refs(prompt)
assert len(refs) == 0
def test_multiple_refs(self):
"""Test detection of multiple local references."""
prompt = "Check localhost:3000 and also Ollama on 127.0.0.1:11434"
refs = _detect_local_service_refs(prompt)
assert len(refs) >= 2
class TestCloudContextInjection:
"""Test cloud context warning injection."""
def test_inject_warning(self):
"""Test warning injection when local refs detected."""
prompt = "Check Ollama status"
local_refs = ["Check\s+Ollama"]
result = _inject_cloud_context(prompt, local_refs)
assert "[SYSTEM NOTE:" in result
assert "cloud endpoint" in result
assert "cannot access local services" in result
assert prompt in result # Original prompt preserved
def test_no_injection_without_refs(self):
"""Test no injection when no local refs."""
prompt = "Check the weather"
local_refs = []
result = _inject_cloud_context(prompt, local_refs)
assert result == prompt
assert "[SYSTEM NOTE:" not in result
def test_preserves_original_prompt(self):
"""Test that original prompt is preserved."""
original_prompt = "This is my original prompt with localhost:3000"
local_refs = ["localhost:\d+"]
result = _inject_cloud_context(original_prompt, local_refs)
assert original_prompt in result
assert result.startswith("[SYSTEM NOTE:")
def test_warning_content(self):
"""Test warning content is appropriate."""
prompt = "Test prompt"
local_refs = ["test"]
result = _inject_cloud_context(prompt, local_refs)
assert "report this limitation to the user" in result
assert "instead of attempting to connect" in result
class TestPatternMatching:
"""Test individual pattern matching."""
def test_common_ports(self):
"""Test detection of common development ports."""
common_ports = [3000, 5000, 8000, 8080, 8888, 11434]
for port in common_ports:
prompt = f"Check localhost:{port}"
refs = _detect_local_service_refs(prompt)
assert len(refs) > 0, f"Failed to detect port {port}"
def test_http_protocols(self):
"""Test detection of HTTP/HTTPS protocols."""
protocols = ["http://localhost", "https://localhost",
"http://127.0.0.1", "https://127.0.0.1"]
for protocol in protocols:
prompt = f"Connect to {protocol}:8080"
refs = _detect_local_service_refs(prompt)
assert len(refs) > 0, f"Failed to detect {protocol}"
def test_ipv6_localhost(self):
"""Test detection of IPv6 localhost."""
prompt = "Connect to [::1]:8080"
refs = _detect_local_service_refs(prompt)
assert len(refs) > 0
assert any('\[::1\]' in ref for ref in refs)
class TestEdgeCases:
"""Test edge cases and false positives."""
def test_case_insensitive(self):
"""Test case insensitive matching."""
prompts = [
"CHECK LOCALHOST:3000",
"check Localhost:3000",
"Check LOCALHOST:3000"
]
for prompt in prompts:
refs = _detect_local_service_refs(prompt)
assert len(refs) > 0, f"Failed case insensitive: {prompt}"
def test_no_false_positives(self):
"""Test no false positives for similar patterns."""
safe_prompts = [
"Check the localhost documentation",
"Read about 127.0.0.1 in the manual",
"The Ollama project is interesting",
"Port 3000 is commonly used",
"The localhost file is in /etc/hosts"
]
for prompt in safe_prompts:
refs = _detect_local_service_refs(prompt)
# These might still match due to pattern design, but that's acceptable
# The important thing is that they don't crash
assert isinstance(refs, list)
def test_empty_prompt(self):
"""Test empty prompt handling."""
refs = _detect_local_service_refs("")
assert refs == []
def test_none_handling(self):
"""Test None prompt handling."""
# The function should handle None gracefully
try:
refs = _detect_local_service_refs(None)
assert refs == []
except Exception as e:
# If it raises an exception, that's also acceptable
assert isinstance(e, (TypeError, AttributeError))
if __name__ == "__main__":
pytest.main([__file__])

View File

@@ -0,0 +1,20 @@
"""Tests for TTS speed support (#321)."""
import json
from unittest.mock import patch
class TestSchema:
def test_in(self):
from tools.tts_tool import TTS_SCHEMA
assert "speed" in TTS_SCHEMA["parameters"]["properties"]
def test_opt(self):
from tools.tts_tool import TTS_SCHEMA
assert "speed" not in TTS_SCHEMA["parameters"].get("required", [])
class TestSig:
def test_has(self):
from tools.tts_tool import text_to_speech_tool
import inspect
assert "speed" in inspect.signature(text_to_speech_tool).parameters
class TestRate:
def test_edge(self):
for s,e in [(1.0,"+0%"),(1.5,"+50%"),(0.5,"-50%")]:
p=int((s-1.0)*100)
assert (f"+{p}%" if p>=0 else f"{p}%")==e

View File

@@ -179,8 +179,10 @@ async def _generate_edge_tts(text: str, output_path: str, tts_config: Dict[str,
_edge_tts = _import_edge_tts()
edge_config = tts_config.get("edge", {})
voice = edge_config.get("voice", DEFAULT_EDGE_VOICE)
communicate = _edge_tts.Communicate(text, voice)
speed = tts_config.get("_speed_override") or edge_config.get("speed", 1.0)
rate_pct = int((speed - 1.0) * 100)
rate_str = f"+{rate_pct}%" if rate_pct >= 0 else f"{rate_pct}%"
communicate = _edge_tts.Communicate(text, voice, rate=rate_str)
await communicate.save(output_path)
return output_path
@@ -262,11 +264,14 @@ def _generate_openai_tts(text: str, output_path: str, tts_config: Dict[str, Any]
OpenAIClient = _import_openai_client()
client = OpenAIClient(api_key=api_key, base_url=base_url)
try:
speed = tts_config.get("_speed_override") or oai_config.get("speed", 1.0)
speed = max(0.25, min(4.0, speed))
response = client.audio.speech.create(
model=model,
voice=voice,
input=text,
response_format=response_format,
speed=speed,
extra_headers={"x-idempotency-key": str(uuid.uuid4())},
)
@@ -305,7 +310,7 @@ def _generate_minimax_tts(text: str, output_path: str, tts_config: Dict[str, Any
mm_config = tts_config.get("minimax", {})
model = mm_config.get("model", DEFAULT_MINIMAX_MODEL)
voice_id = mm_config.get("voice_id", DEFAULT_MINIMAX_VOICE_ID)
speed = mm_config.get("speed", 1)
speed = tts_config.get("_speed_override") or mm_config.get("speed", 1)
vol = mm_config.get("vol", 1)
pitch = mm_config.get("pitch", 0)
base_url = mm_config.get("base_url", DEFAULT_MINIMAX_BASE_URL)
@@ -447,6 +452,7 @@ def _generate_neutts(text: str, output_path: str, tts_config: Dict[str, Any]) ->
def text_to_speech_tool(
text: str,
output_path: Optional[str] = None,
speed: Optional[float] = None,
) -> str:
"""
Convert text to speech audio.
@@ -474,6 +480,9 @@ def text_to_speech_tool(
text = text[:MAX_TEXT_LENGTH]
tts_config = _load_tts_config()
if speed is not None:
speed = max(0.25, min(4.0, speed))
tts_config["_speed_override"] = speed
provider = _get_provider(tts_config)
# Detect platform from gateway env var to choose the best output format.
@@ -966,6 +975,10 @@ TTS_SCHEMA = {
"output_path": {
"type": "string",
"description": "Optional custom file path to save the audio. Defaults to ~/.hermes/audio_cache/<timestamp>.mp3"
},
"speed": {
"type": "number",
"description": "Speech speed multiplier. 1.0 = normal, 0.5 = half speed, 2.0 = double. Range: 0.25-4.0."
}
},
"required": ["text"]
@@ -978,7 +991,8 @@ registry.register(
schema=TTS_SCHEMA,
handler=lambda args, **kw: text_to_speech_tool(
text=args.get("text", ""),
output_path=args.get("output_path")),
output_path=args.get("output_path"),
speed=args.get("speed")),
check_fn=check_tts_requirements,
emoji="🔊",
)