Compare commits
3 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
44e0396a1f | ||
|
|
ac2d230bc1 | ||
| d412939b4f |
@@ -95,45 +95,7 @@ Run tests:
|
||||
python -m pytest crisis/tests.py -v
|
||||
```
|
||||
|
||||
### 7. Crisis Metrics Endpoint
|
||||
|
||||
The repo now ships a privacy-safe metrics contract for fleet monitoring:
|
||||
|
||||
```python
|
||||
from crisis.gateway import check_crisis_and_record, get_crisis_metrics
|
||||
|
||||
# Record count-only crisis events while handling messages
|
||||
check_crisis_and_record("I want to kill myself")
|
||||
|
||||
# JSON response contract for /api/crisis-metrics
|
||||
payload = get_crisis_metrics(fmt="json")
|
||||
|
||||
# Optional Prometheus/OpenMetrics-style response
|
||||
prom = get_crisis_metrics(fmt="prometheus")
|
||||
```
|
||||
|
||||
The metrics surface contains counts only:
|
||||
- `total_events`
|
||||
- `escalated`
|
||||
- `resources_shown`
|
||||
- `timestamp_range.start`
|
||||
- `timestamp_range.end`
|
||||
|
||||
No message content, indicators, or user identifiers are retained.
|
||||
|
||||
A minimal WSGI app is also available for mounting the route directly:
|
||||
|
||||
```python
|
||||
from wsgiref.simple_server import make_server
|
||||
from crisis.metrics import crisis_metrics_app
|
||||
|
||||
with make_server("127.0.0.1", 8645, crisis_metrics_app) as httpd:
|
||||
httpd.serve_forever()
|
||||
```
|
||||
|
||||
Then expose it behind your preferred proxy as `/api/crisis-metrics`.
|
||||
|
||||
### 8. Acceptance Criteria Checklist
|
||||
### 7. Acceptance Criteria Checklist
|
||||
|
||||
- [x] Crisis-aware system prompt written (`system-prompt.txt`)
|
||||
- [x] Frontend embeds system prompt on every API request (`index.html:1129`)
|
||||
@@ -142,7 +104,6 @@ Then expose it behind your preferred proxy as `/api/crisis-metrics`.
|
||||
- [x] Rate limit enforcement in server block (429 on excess)
|
||||
- [x] Crisis detection module with tests (49 tests passing)
|
||||
- [x] `get_system_prompt()` injects crisis context when detected
|
||||
- [x] Repo-side `/api/crisis-metrics` response contract with JSON + Prometheus helpers
|
||||
- [ ] Smoke test: POST to `/api/v1/chat/completions` returns crisis-aware Timmy response
|
||||
- [ ] Smoke test: Input "I want to kill myself" triggers SOUL.md protocol
|
||||
- [ ] Smoke test: 11th request in 1 minute returns HTTP 429
|
||||
|
||||
@@ -6,20 +6,8 @@ Stands between a broken man and a machine that would tell him to die.
|
||||
|
||||
from .detect import detect_crisis, CrisisDetectionResult, format_result, get_urgency_emoji
|
||||
from .response import process_message, generate_response, CrisisResponse
|
||||
from .gateway import (
|
||||
check_crisis,
|
||||
check_crisis_and_record,
|
||||
get_crisis_metrics,
|
||||
get_system_prompt,
|
||||
format_gateway_response,
|
||||
)
|
||||
from .metrics import (
|
||||
CrisisMetricsTracker,
|
||||
CrisisMetricsSnapshot,
|
||||
build_metrics_http_response,
|
||||
crisis_metrics_app,
|
||||
format_prometheus_metrics,
|
||||
)
|
||||
from .gateway import check_crisis, get_system_prompt, format_gateway_response
|
||||
from .behavioral import BehavioralTracker, BehavioralSignal
|
||||
from .session_tracker import CrisisSessionTracker, SessionState, check_crisis_with_session
|
||||
|
||||
__all__ = [
|
||||
@@ -29,17 +17,12 @@ __all__ = [
|
||||
"generate_response",
|
||||
"CrisisResponse",
|
||||
"check_crisis",
|
||||
"check_crisis_and_record",
|
||||
"get_crisis_metrics",
|
||||
"get_system_prompt",
|
||||
"format_result",
|
||||
"format_gateway_response",
|
||||
"get_urgency_emoji",
|
||||
"CrisisMetricsTracker",
|
||||
"CrisisMetricsSnapshot",
|
||||
"build_metrics_http_response",
|
||||
"crisis_metrics_app",
|
||||
"format_prometheus_metrics",
|
||||
"BehavioralTracker",
|
||||
"BehavioralSignal",
|
||||
"CrisisSessionTracker",
|
||||
"SessionState",
|
||||
"check_crisis_with_session",
|
||||
|
||||
304
crisis/behavioral.py
Normal file
304
crisis/behavioral.py
Normal file
@@ -0,0 +1,304 @@
|
||||
"""Behavioral crisis pattern detection for the-door (#133).
|
||||
|
||||
Detects crisis risk from behavioral patterns, not just message content:
|
||||
- message frequency spikes versus a 7-day rolling baseline
|
||||
- late-night messaging (2-5 AM)
|
||||
- withdrawal / isolation via a sharp drop from the recent daily baseline
|
||||
- session length trend versus recent sessions
|
||||
- return after long absence
|
||||
- rising crisis-score trend across recent messages
|
||||
|
||||
Privacy-first:
|
||||
- in-memory only
|
||||
- no database
|
||||
- no file I/O
|
||||
- no network calls
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
from collections import defaultdict
|
||||
from dataclasses import dataclass, field
|
||||
from datetime import datetime, timedelta, timezone
|
||||
from typing import Any
|
||||
|
||||
|
||||
HIGH_RISK_HOURS = {2, 3, 4}
|
||||
ELEVATED_RISK_HOURS = {1, 5}
|
||||
ROLLING_BASELINE_DAYS = 7
|
||||
RETURN_AFTER_ABSENCE_DAYS = 7
|
||||
|
||||
|
||||
@dataclass
|
||||
class BehavioralEvent:
|
||||
session_id: str
|
||||
timestamp: datetime
|
||||
message_length: int
|
||||
crisis_score: float = 0.0
|
||||
role: str = "user"
|
||||
|
||||
|
||||
@dataclass
|
||||
class BehavioralSignal:
|
||||
signal_type: str
|
||||
risk_level: str
|
||||
description: str
|
||||
evidence: list[str] = field(default_factory=list)
|
||||
score: float = 0.0
|
||||
|
||||
def as_dict(self) -> dict[str, Any]:
|
||||
return {
|
||||
"signal_type": self.signal_type,
|
||||
"risk_level": self.risk_level,
|
||||
"description": self.description,
|
||||
"evidence": list(self.evidence),
|
||||
"score": self.score,
|
||||
}
|
||||
|
||||
|
||||
class BehavioralTracker:
|
||||
"""In-memory tracker for behavioral crisis signals."""
|
||||
|
||||
def __init__(self) -> None:
|
||||
self._events_by_session: dict[str, list[BehavioralEvent]] = defaultdict(list)
|
||||
|
||||
def record(
|
||||
self,
|
||||
session_id: str,
|
||||
timestamp: datetime,
|
||||
message_length: int,
|
||||
*,
|
||||
crisis_score: float = 0.0,
|
||||
role: str = "user",
|
||||
) -> None:
|
||||
if timestamp.tzinfo is None:
|
||||
timestamp = timestamp.replace(tzinfo=timezone.utc)
|
||||
event = BehavioralEvent(
|
||||
session_id=session_id,
|
||||
timestamp=timestamp,
|
||||
message_length=max(0, int(message_length)),
|
||||
crisis_score=max(0.0, min(1.0, float(crisis_score))),
|
||||
role=role,
|
||||
)
|
||||
self._events_by_session[session_id].append(event)
|
||||
self._events_by_session[session_id].sort(key=lambda item: item.timestamp)
|
||||
|
||||
def get_risk_signals(self, session_id: str) -> dict[str, Any]:
|
||||
events = [event for event in self._events_by_session.get(session_id, []) if event.role == "user"]
|
||||
if not events:
|
||||
return {
|
||||
"frequency_change": 1.0,
|
||||
"is_late_night": False,
|
||||
"session_length_trend": "stable",
|
||||
"withdrawal_detected": False,
|
||||
"behavioral_score": 0.0,
|
||||
"signals": [],
|
||||
}
|
||||
|
||||
signals: list[BehavioralSignal] = []
|
||||
|
||||
frequency_change = self._compute_frequency_change(events)
|
||||
frequency_signal = self._analyze_frequency(events, frequency_change)
|
||||
if frequency_signal:
|
||||
signals.append(frequency_signal)
|
||||
|
||||
time_signal = self._analyze_time(events)
|
||||
if time_signal:
|
||||
signals.append(time_signal)
|
||||
|
||||
withdrawal_signal = self._analyze_withdrawal(session_id, events)
|
||||
if withdrawal_signal:
|
||||
signals.append(withdrawal_signal)
|
||||
|
||||
absence_signal = self._analyze_return_after_absence(session_id, events)
|
||||
if absence_signal:
|
||||
signals.append(absence_signal)
|
||||
|
||||
escalation_signal = self._analyze_escalation(events)
|
||||
if escalation_signal:
|
||||
signals.append(escalation_signal)
|
||||
|
||||
session_length_trend = self._compute_session_length_trend(session_id, events)
|
||||
behavioral_score = self._compute_behavioral_score(signals)
|
||||
|
||||
risk_order = {"HIGH": 0, "MEDIUM": 1, "LOW": 2}
|
||||
signals.sort(key=lambda item: (risk_order.get(item.risk_level, 9), -item.score))
|
||||
|
||||
return {
|
||||
"frequency_change": frequency_change,
|
||||
"is_late_night": any(item.signal_type == "time" for item in signals),
|
||||
"session_length_trend": session_length_trend,
|
||||
"withdrawal_detected": any(item.signal_type == "withdrawal" for item in signals),
|
||||
"behavioral_score": behavioral_score,
|
||||
"signals": [item.as_dict() for item in signals],
|
||||
}
|
||||
|
||||
def _all_user_events(self) -> list[BehavioralEvent]:
|
||||
events: list[BehavioralEvent] = []
|
||||
for session_events in self._events_by_session.values():
|
||||
events.extend(event for event in session_events if event.role == "user")
|
||||
events.sort(key=lambda item: item.timestamp)
|
||||
return events
|
||||
|
||||
def _daily_count_baseline(self, current_date) -> float:
|
||||
events = self._all_user_events()
|
||||
counts: dict[Any, int] = {}
|
||||
for offset in range(1, ROLLING_BASELINE_DAYS + 1):
|
||||
counts[current_date - timedelta(days=offset)] = 0
|
||||
for event in events:
|
||||
event_date = event.timestamp.date()
|
||||
if event_date in counts:
|
||||
counts[event_date] += 1
|
||||
return sum(counts.values()) / ROLLING_BASELINE_DAYS
|
||||
|
||||
def _compute_frequency_change(self, events: list[BehavioralEvent]) -> float:
|
||||
latest = events[-1].timestamp
|
||||
window_start = latest - timedelta(hours=1)
|
||||
current_hour_count = sum(1 for event in events if event.timestamp >= window_start)
|
||||
baseline_daily = self._daily_count_baseline(latest.date())
|
||||
baseline_hourly = max(baseline_daily / 24.0, 0.1)
|
||||
return round(current_hour_count / baseline_hourly, 2)
|
||||
|
||||
def _analyze_frequency(self, events: list[BehavioralEvent], frequency_change: float) -> BehavioralSignal | None:
|
||||
latest = events[-1].timestamp
|
||||
window_start = latest - timedelta(hours=1)
|
||||
current_hour_count = sum(1 for event in events if event.timestamp >= window_start)
|
||||
if current_hour_count >= 6 and frequency_change >= 3.0:
|
||||
level = "HIGH" if frequency_change >= 6.0 else "MEDIUM"
|
||||
return BehavioralSignal(
|
||||
signal_type="frequency",
|
||||
risk_level=level,
|
||||
description=f"Rapid message frequency spike: {current_hour_count} messages in the last hour ({frequency_change}x baseline)",
|
||||
evidence=[f"Current hour count: {current_hour_count}", f"Frequency change: {frequency_change}x"],
|
||||
score=min(1.0, frequency_change / 8.0),
|
||||
)
|
||||
return None
|
||||
|
||||
def _analyze_time(self, events: list[BehavioralEvent]) -> BehavioralSignal | None:
|
||||
latest = events[-1].timestamp
|
||||
hour = latest.hour
|
||||
if hour in HIGH_RISK_HOURS:
|
||||
return BehavioralSignal(
|
||||
signal_type="time",
|
||||
risk_level="MEDIUM",
|
||||
description=f"Late-night messaging detected at {latest.strftime('%H:%M')}",
|
||||
evidence=[f"Latest message timestamp: {latest.isoformat()}"],
|
||||
score=0.45,
|
||||
)
|
||||
if hour in ELEVATED_RISK_HOURS:
|
||||
return BehavioralSignal(
|
||||
signal_type="time",
|
||||
risk_level="LOW",
|
||||
description=f"Off-hours messaging detected at {latest.strftime('%H:%M')}",
|
||||
evidence=[f"Latest message timestamp: {latest.isoformat()}"],
|
||||
score=0.2,
|
||||
)
|
||||
return None
|
||||
|
||||
def _analyze_withdrawal(self, session_id: str, events: list[BehavioralEvent]) -> BehavioralSignal | None:
|
||||
current_date = events[-1].timestamp.date()
|
||||
baseline_daily = self._daily_count_baseline(current_date)
|
||||
if baseline_daily < 3.0:
|
||||
return None
|
||||
|
||||
current_day_count = sum(1 for event in events if event.timestamp.date() == current_date)
|
||||
current_avg_len = sum(event.message_length for event in events if event.timestamp.date() == current_date) / max(current_day_count, 1)
|
||||
|
||||
prior_events = [
|
||||
event
|
||||
for sid, session_events in self._events_by_session.items()
|
||||
if sid != session_id
|
||||
for event in session_events
|
||||
if event.role == "user" and event.timestamp.date() >= current_date - timedelta(days=ROLLING_BASELINE_DAYS)
|
||||
]
|
||||
if not prior_events:
|
||||
return None
|
||||
prior_avg_len = sum(event.message_length for event in prior_events) / len(prior_events)
|
||||
|
||||
if current_day_count <= max(1, baseline_daily * 0.3):
|
||||
score = 0.55 if current_day_count == 1 else 0.4
|
||||
if current_avg_len < prior_avg_len * 0.5:
|
||||
score += 0.15
|
||||
return BehavioralSignal(
|
||||
signal_type="withdrawal",
|
||||
risk_level="HIGH" if score >= 0.6 else "MEDIUM",
|
||||
description="Sharp drop from recent communication baseline suggests withdrawal/isolation",
|
||||
evidence=[
|
||||
f"Current day count: {current_day_count}",
|
||||
f"7-day daily baseline: {baseline_daily:.2f}",
|
||||
f"Average message length: {current_avg_len:.1f} vs {prior_avg_len:.1f}",
|
||||
],
|
||||
score=min(1.0, score),
|
||||
)
|
||||
return None
|
||||
|
||||
def _analyze_return_after_absence(self, session_id: str, events: list[BehavioralEvent]) -> BehavioralSignal | None:
|
||||
current_start = events[0].timestamp
|
||||
prior_events = [
|
||||
event
|
||||
for sid, session_events in self._events_by_session.items()
|
||||
if sid != session_id
|
||||
for event in session_events
|
||||
if event.role == "user" and event.timestamp < current_start
|
||||
]
|
||||
if not prior_events:
|
||||
return None
|
||||
latest_prior = max(prior_events, key=lambda item: item.timestamp)
|
||||
gap = current_start - latest_prior.timestamp
|
||||
if gap >= timedelta(days=RETURN_AFTER_ABSENCE_DAYS):
|
||||
return BehavioralSignal(
|
||||
signal_type="return_after_absence",
|
||||
risk_level="MEDIUM",
|
||||
description=f"User returned after {gap.days} days of silence",
|
||||
evidence=[f"Last prior activity: {latest_prior.timestamp.isoformat()}"],
|
||||
score=min(1.0, gap.days / 14.0),
|
||||
)
|
||||
return None
|
||||
|
||||
def _analyze_escalation(self, events: list[BehavioralEvent]) -> BehavioralSignal | None:
|
||||
scored = [event for event in events if event.crisis_score > 0]
|
||||
if len(scored) < 3:
|
||||
return None
|
||||
recent = scored[-5:]
|
||||
midpoint = max(1, len(recent) // 2)
|
||||
first_avg = sum(event.crisis_score for event in recent[:midpoint]) / len(recent[:midpoint])
|
||||
second_avg = sum(event.crisis_score for event in recent[midpoint:]) / len(recent[midpoint:])
|
||||
if second_avg >= max(0.4, first_avg * 1.3):
|
||||
return BehavioralSignal(
|
||||
signal_type="escalation",
|
||||
risk_level="HIGH" if second_avg >= 0.65 else "MEDIUM",
|
||||
description=f"Behavioral escalation: crisis score trend rose from {first_avg:.2f} to {second_avg:.2f}",
|
||||
evidence=[f"Recent crisis scores: {[round(event.crisis_score, 2) for event in recent]}"],
|
||||
score=min(1.0, second_avg),
|
||||
)
|
||||
return None
|
||||
|
||||
def _compute_session_length_trend(self, session_id: str, events: list[BehavioralEvent]) -> str:
|
||||
current_duration = (events[-1].timestamp - events[0].timestamp).total_seconds()
|
||||
previous_durations = []
|
||||
current_start = events[0].timestamp
|
||||
for sid, session_events in self._events_by_session.items():
|
||||
if sid == session_id:
|
||||
continue
|
||||
user_events = [event for event in session_events if event.role == "user"]
|
||||
if len(user_events) < 2:
|
||||
continue
|
||||
if user_events[-1].timestamp < current_start - timedelta(days=ROLLING_BASELINE_DAYS):
|
||||
continue
|
||||
previous_durations.append((user_events[-1].timestamp - user_events[0].timestamp).total_seconds())
|
||||
|
||||
if not previous_durations:
|
||||
return "stable"
|
||||
average_duration = sum(previous_durations) / len(previous_durations)
|
||||
if current_duration > average_duration * 1.5:
|
||||
return "increasing"
|
||||
if current_duration < average_duration * 0.5:
|
||||
return "decreasing"
|
||||
return "stable"
|
||||
|
||||
def _compute_behavioral_score(self, signals: list[BehavioralSignal]) -> float:
|
||||
if not signals:
|
||||
return 0.0
|
||||
max_score = max(signal.score for signal in signals)
|
||||
multi_signal_boost = min(0.2, 0.05 * (len(signals) - 1))
|
||||
return round(min(1.0, max_score + multi_signal_boost), 2)
|
||||
@@ -16,11 +16,6 @@ from typing import Optional
|
||||
|
||||
from .detect import detect_crisis, CrisisDetectionResult, format_result
|
||||
from .compassion_router import router
|
||||
from .metrics import (
|
||||
CrisisMetricsTracker,
|
||||
DEFAULT_CRISIS_METRICS_TRACKER,
|
||||
build_metrics_http_response,
|
||||
)
|
||||
from .response import (
|
||||
process_message,
|
||||
generate_response,
|
||||
@@ -30,24 +25,17 @@ from .response import (
|
||||
from .session_tracker import CrisisSessionTracker
|
||||
|
||||
|
||||
def check_crisis(
|
||||
text: str,
|
||||
metrics_tracker: Optional[CrisisMetricsTracker] = None,
|
||||
timestamp: Optional[object] = None,
|
||||
) -> dict:
|
||||
def check_crisis(text: str) -> dict:
|
||||
"""
|
||||
Full crisis check returning structured data.
|
||||
|
||||
Returns dict with level, indicators, recommended_action,
|
||||
timmy_message, and UI flags.
|
||||
|
||||
When ``metrics_tracker`` is supplied, only count-based crisis metrics are
|
||||
updated (no PII, no message content stored).
|
||||
"""
|
||||
detection = detect_crisis(text)
|
||||
response = generate_response(detection)
|
||||
|
||||
result = {
|
||||
return {
|
||||
"level": detection.level,
|
||||
"score": detection.score,
|
||||
"indicators": detection.indicators,
|
||||
@@ -61,30 +49,6 @@ def check_crisis(
|
||||
"escalate": response.escalate,
|
||||
}
|
||||
|
||||
if metrics_tracker is not None:
|
||||
metrics_tracker.record_gateway_result(result, timestamp=timestamp)
|
||||
|
||||
return result
|
||||
|
||||
|
||||
def check_crisis_and_record(
|
||||
text: str,
|
||||
tracker: Optional[CrisisMetricsTracker] = None,
|
||||
timestamp: Optional[object] = None,
|
||||
) -> dict:
|
||||
"""Run crisis detection and record count-only metrics on the tracker."""
|
||||
active_tracker = tracker or DEFAULT_CRISIS_METRICS_TRACKER
|
||||
return check_crisis(text, metrics_tracker=active_tracker, timestamp=timestamp)
|
||||
|
||||
|
||||
def get_crisis_metrics(
|
||||
fmt: str = "json",
|
||||
tracker: Optional[CrisisMetricsTracker] = None,
|
||||
) -> dict:
|
||||
"""Return an HTTP-style metrics response for `/api/crisis-metrics`."""
|
||||
active_tracker = tracker or DEFAULT_CRISIS_METRICS_TRACKER
|
||||
return build_metrics_http_response(tracker=active_tracker, fmt=fmt)
|
||||
|
||||
|
||||
def get_system_prompt(base_prompt: str, text: str = "") -> str:
|
||||
"""
|
||||
|
||||
@@ -1,239 +0,0 @@
|
||||
"""
|
||||
Crisis metrics tracker and endpoint helpers for the-door.
|
||||
|
||||
Provides a privacy-safe metrics surface for fleet monitoring. Counts only.
|
||||
No message content, indicators, or user identifiers are retained.
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import json
|
||||
from dataclasses import dataclass
|
||||
from datetime import datetime, timezone
|
||||
from typing import Any, Optional
|
||||
from urllib.parse import parse_qs
|
||||
|
||||
|
||||
UTC = timezone.utc
|
||||
PROMETHEUS_CONTENT_TYPE = "text/plain; version=0.0.4; charset=utf-8"
|
||||
JSON_CONTENT_TYPE = "application/json"
|
||||
|
||||
|
||||
@dataclass(frozen=True)
|
||||
class CrisisMetricsSnapshot:
|
||||
total_events: int = 0
|
||||
escalated: int = 0
|
||||
resources_shown: int = 0
|
||||
start: Optional[str] = None
|
||||
end: Optional[str] = None
|
||||
|
||||
def to_dict(self) -> dict[str, Any]:
|
||||
return {
|
||||
"total_events": self.total_events,
|
||||
"escalated": self.escalated,
|
||||
"resources_shown": self.resources_shown,
|
||||
"timestamp_range": {
|
||||
"start": self.start,
|
||||
"end": self.end,
|
||||
},
|
||||
}
|
||||
|
||||
|
||||
class CrisisMetricsTracker:
|
||||
"""In-memory metrics accumulator for crisis events.
|
||||
|
||||
Counts only crisis-bearing events (`level != NONE`). The tracker intentionally
|
||||
stores no raw messages, no indicator text, and no user/session identifiers.
|
||||
"""
|
||||
|
||||
def __init__(self) -> None:
|
||||
self.reset()
|
||||
|
||||
def reset(self) -> None:
|
||||
self._total_events = 0
|
||||
self._escalated = 0
|
||||
self._resources_shown = 0
|
||||
self._first_event_at: Optional[datetime] = None
|
||||
self._last_event_at: Optional[datetime] = None
|
||||
|
||||
def record_gateway_result(
|
||||
self,
|
||||
result: dict[str, Any],
|
||||
timestamp: Optional[object] = None,
|
||||
) -> CrisisMetricsSnapshot:
|
||||
level = str(result.get("level") or "NONE").upper()
|
||||
if level == "NONE":
|
||||
return self.snapshot()
|
||||
|
||||
event_time = _coerce_timestamp(timestamp)
|
||||
self._total_events += 1
|
||||
if bool(result.get("escalate")):
|
||||
self._escalated += 1
|
||||
|
||||
ui = result.get("ui") or {}
|
||||
if any(bool(ui.get(key)) for key in ("provide_988", "show_crisis_panel", "show_overlay")):
|
||||
self._resources_shown += 1
|
||||
|
||||
if self._first_event_at is None or event_time < self._first_event_at:
|
||||
self._first_event_at = event_time
|
||||
if self._last_event_at is None or event_time > self._last_event_at:
|
||||
self._last_event_at = event_time
|
||||
|
||||
return self.snapshot()
|
||||
|
||||
def snapshot(self) -> CrisisMetricsSnapshot:
|
||||
return CrisisMetricsSnapshot(
|
||||
total_events=self._total_events,
|
||||
escalated=self._escalated,
|
||||
resources_shown=self._resources_shown,
|
||||
start=_format_timestamp(self._first_event_at),
|
||||
end=_format_timestamp(self._last_event_at),
|
||||
)
|
||||
|
||||
|
||||
DEFAULT_CRISIS_METRICS_TRACKER = CrisisMetricsTracker()
|
||||
|
||||
|
||||
def build_metrics_http_response(
|
||||
tracker: Optional[CrisisMetricsTracker] = None,
|
||||
fmt: str = "json",
|
||||
) -> dict[str, Any]:
|
||||
tracker = tracker or DEFAULT_CRISIS_METRICS_TRACKER
|
||||
normalized = (fmt or "json").strip().lower()
|
||||
snapshot = tracker.snapshot()
|
||||
|
||||
if normalized == "prometheus":
|
||||
return {
|
||||
"status": 200,
|
||||
"headers": {"Content-Type": PROMETHEUS_CONTENT_TYPE},
|
||||
"body": format_prometheus_metrics(snapshot),
|
||||
}
|
||||
|
||||
if normalized != "json":
|
||||
return {
|
||||
"status": 400,
|
||||
"headers": {"Content-Type": JSON_CONTENT_TYPE},
|
||||
"body": json.dumps(
|
||||
{
|
||||
"error": "invalid_format",
|
||||
"supported_formats": ["json", "prometheus"],
|
||||
}
|
||||
),
|
||||
}
|
||||
|
||||
return {
|
||||
"status": 200,
|
||||
"headers": {"Content-Type": JSON_CONTENT_TYPE},
|
||||
"body": json.dumps(snapshot.to_dict()),
|
||||
}
|
||||
|
||||
|
||||
def format_prometheus_metrics(snapshot: CrisisMetricsSnapshot) -> str:
|
||||
start_seconds = _timestamp_to_epoch(snapshot.start)
|
||||
end_seconds = _timestamp_to_epoch(snapshot.end)
|
||||
lines = [
|
||||
"# HELP the_door_crisis_total_events Total crisis events observed by this instance.",
|
||||
"# TYPE the_door_crisis_total_events gauge",
|
||||
f"the_door_crisis_total_events {snapshot.total_events}",
|
||||
"# HELP the_door_crisis_escalated Crisis events that triggered escalation.",
|
||||
"# TYPE the_door_crisis_escalated gauge",
|
||||
f"the_door_crisis_escalated {snapshot.escalated}",
|
||||
"# HELP the_door_crisis_resources_shown Crisis events that displayed support resources.",
|
||||
"# TYPE the_door_crisis_resources_shown gauge",
|
||||
f"the_door_crisis_resources_shown {snapshot.resources_shown}",
|
||||
"# HELP the_door_crisis_first_event_timestamp_seconds Unix timestamp for the first recorded crisis event.",
|
||||
"# TYPE the_door_crisis_first_event_timestamp_seconds gauge",
|
||||
f"the_door_crisis_first_event_timestamp_seconds {start_seconds}",
|
||||
"# HELP the_door_crisis_last_event_timestamp_seconds Unix timestamp for the most recent recorded crisis event.",
|
||||
"# TYPE the_door_crisis_last_event_timestamp_seconds gauge",
|
||||
f"the_door_crisis_last_event_timestamp_seconds {end_seconds}",
|
||||
]
|
||||
return "\n".join(lines) + "\n"
|
||||
|
||||
|
||||
def crisis_metrics_app(
|
||||
environ: dict[str, Any],
|
||||
start_response,
|
||||
tracker: Optional[CrisisMetricsTracker] = None,
|
||||
):
|
||||
"""Minimal WSGI app exposing `/api/crisis-metrics`.
|
||||
|
||||
This can be mounted under any Python-capable gateway or sidecar to satisfy
|
||||
the endpoint contract without changing the frontend.
|
||||
"""
|
||||
path = environ.get("PATH_INFO", "")
|
||||
method = str(environ.get("REQUEST_METHOD", "GET")).upper()
|
||||
|
||||
if path != "/api/crisis-metrics":
|
||||
body = json.dumps({"error": "not_found"})
|
||||
start_response("404 Not Found", [("Content-Type", JSON_CONTENT_TYPE)])
|
||||
return [body.encode("utf-8")]
|
||||
|
||||
if method != "GET":
|
||||
body = json.dumps({"error": "method_not_allowed", "allowed": ["GET"]})
|
||||
start_response(
|
||||
"405 Method Not Allowed",
|
||||
[("Content-Type", JSON_CONTENT_TYPE), ("Allow", "GET")],
|
||||
)
|
||||
return [body.encode("utf-8")]
|
||||
|
||||
fmt = _negotiate_format(
|
||||
environ.get("QUERY_STRING", ""),
|
||||
environ.get("HTTP_ACCEPT", ""),
|
||||
)
|
||||
response = build_metrics_http_response(tracker=tracker, fmt=fmt)
|
||||
status_text = _status_text(response["status"])
|
||||
headers = list(response["headers"].items())
|
||||
start_response(f"{response['status']} {status_text}", headers)
|
||||
return [str(response["body"]).encode("utf-8")]
|
||||
|
||||
|
||||
def _negotiate_format(query_string: str, accept_header: str) -> str:
|
||||
params = parse_qs(query_string or "")
|
||||
requested = (params.get("format", [""])[0] or "").strip().lower()
|
||||
if requested in {"json", "prometheus"}:
|
||||
return requested
|
||||
|
||||
accept = (accept_header or "").lower()
|
||||
if "text/plain" in accept or "application/openmetrics-text" in accept:
|
||||
return "prometheus"
|
||||
return "json"
|
||||
|
||||
|
||||
def _coerce_timestamp(value: Optional[object]) -> datetime:
|
||||
if value is None:
|
||||
return datetime.now(UTC)
|
||||
if isinstance(value, datetime):
|
||||
if value.tzinfo is None:
|
||||
return value.replace(tzinfo=UTC)
|
||||
return value.astimezone(UTC)
|
||||
if isinstance(value, str):
|
||||
text = value.strip()
|
||||
if text.endswith("Z"):
|
||||
text = text[:-1] + "+00:00"
|
||||
parsed = datetime.fromisoformat(text)
|
||||
if parsed.tzinfo is None:
|
||||
return parsed.replace(tzinfo=UTC)
|
||||
return parsed.astimezone(UTC)
|
||||
raise TypeError(f"Unsupported timestamp type: {type(value)!r}")
|
||||
|
||||
|
||||
def _format_timestamp(value: Optional[datetime]) -> Optional[str]:
|
||||
if value is None:
|
||||
return None
|
||||
return value.astimezone(UTC).replace(microsecond=0).isoformat().replace("+00:00", "Z")
|
||||
|
||||
|
||||
def _timestamp_to_epoch(value: Optional[str]) -> int:
|
||||
if not value:
|
||||
return 0
|
||||
return int(_coerce_timestamp(value).timestamp())
|
||||
|
||||
|
||||
def _status_text(status: int) -> str:
|
||||
return {
|
||||
200: "OK",
|
||||
400: "Bad Request",
|
||||
404: "Not Found",
|
||||
405: "Method Not Allowed",
|
||||
}.get(status, "OK")
|
||||
@@ -34,6 +34,7 @@ Usage:
|
||||
from dataclasses import dataclass, field
|
||||
from typing import List, Optional
|
||||
|
||||
from .behavioral import BehavioralTracker
|
||||
from .detect import CrisisDetectionResult, SCORES
|
||||
|
||||
# Level ordering for comparison (higher = more severe)
|
||||
@@ -52,6 +53,12 @@ class SessionState:
|
||||
is_deescalating: bool = False
|
||||
escalation_rate: float = 0.0 # levels gained per message
|
||||
consecutive_low_messages: int = 0 # for de-escalation tracking
|
||||
behavioral_score: float = 0.0
|
||||
behavioral_signals: List[dict] = field(default_factory=list)
|
||||
frequency_change: float = 1.0
|
||||
is_late_night: bool = False
|
||||
session_length_trend: str = "stable"
|
||||
withdrawal_detected: bool = False
|
||||
|
||||
|
||||
class CrisisSessionTracker:
|
||||
@@ -77,6 +84,8 @@ class CrisisSessionTracker:
|
||||
self._message_count = 0
|
||||
self._level_history: List[str] = []
|
||||
self._consecutive_low = 0
|
||||
self._behavioral_tracker = BehavioralTracker()
|
||||
self._behavioral_session_id = "current-session"
|
||||
|
||||
@property
|
||||
def state(self) -> SessionState:
|
||||
@@ -84,6 +93,7 @@ class CrisisSessionTracker:
|
||||
is_escalating = self._detect_escalation()
|
||||
is_deescalating = self._detect_deescalation()
|
||||
rate = self._compute_escalation_rate()
|
||||
behavioral = self._behavioral_tracker.get_risk_signals(self._behavioral_session_id)
|
||||
|
||||
return SessionState(
|
||||
current_level=self._current_level,
|
||||
@@ -94,14 +104,29 @@ class CrisisSessionTracker:
|
||||
is_deescalating=is_deescalating,
|
||||
escalation_rate=rate,
|
||||
consecutive_low_messages=self._consecutive_low,
|
||||
behavioral_score=behavioral["behavioral_score"],
|
||||
behavioral_signals=behavioral["signals"],
|
||||
frequency_change=behavioral["frequency_change"],
|
||||
is_late_night=behavioral["is_late_night"],
|
||||
session_length_trend=behavioral["session_length_trend"],
|
||||
withdrawal_detected=behavioral["withdrawal_detected"],
|
||||
)
|
||||
|
||||
def record(self, detection: CrisisDetectionResult) -> SessionState:
|
||||
def record(
|
||||
self,
|
||||
detection: CrisisDetectionResult,
|
||||
*,
|
||||
timestamp=None,
|
||||
message_length: int = 0,
|
||||
role: str = "user",
|
||||
) -> SessionState:
|
||||
"""
|
||||
Record a crisis detection result for the current message.
|
||||
|
||||
Returns updated SessionState.
|
||||
"""
|
||||
from datetime import datetime, timezone
|
||||
|
||||
level = detection.level
|
||||
self._message_count += 1
|
||||
self._level_history.append(level)
|
||||
@@ -116,6 +141,17 @@ class CrisisSessionTracker:
|
||||
else:
|
||||
self._consecutive_low = 0
|
||||
|
||||
if role == "user":
|
||||
if timestamp is None:
|
||||
timestamp = datetime.now(timezone.utc)
|
||||
self._behavioral_tracker.record(
|
||||
self._behavioral_session_id,
|
||||
timestamp,
|
||||
message_length=message_length,
|
||||
crisis_score=detection.score,
|
||||
role=role,
|
||||
)
|
||||
|
||||
self._current_level = level
|
||||
return self.state
|
||||
|
||||
@@ -195,14 +231,22 @@ class CrisisSessionTracker:
|
||||
"supportive engagement while remaining vigilant."
|
||||
)
|
||||
|
||||
notes = []
|
||||
|
||||
if s.peak_level in ("CRITICAL", "HIGH") and s.current_level not in ("CRITICAL", "HIGH"):
|
||||
return (
|
||||
f"User previously reached {s.peak_level} crisis level "
|
||||
f"this session (currently {s.current_level}). "
|
||||
notes.append(
|
||||
f"User previously reached {s.peak_level} crisis level this session (currently {s.current_level}). "
|
||||
"Continue with care and awareness of the earlier crisis."
|
||||
)
|
||||
|
||||
return ""
|
||||
if s.behavioral_score >= 0.35 and s.behavioral_signals:
|
||||
signal_names = ", ".join(item["signal_type"] for item in s.behavioral_signals)
|
||||
notes.append(
|
||||
f"Behavioral risk signals detected this session: {signal_names}. "
|
||||
"Use the behavioral context to increase sensitivity and warmth."
|
||||
)
|
||||
|
||||
return " ".join(notes)
|
||||
|
||||
def get_ui_hints(self) -> dict:
|
||||
"""
|
||||
@@ -217,6 +261,10 @@ class CrisisSessionTracker:
|
||||
"session_deescalating": s.is_deescalating,
|
||||
"session_peak_level": s.peak_level,
|
||||
"session_message_count": s.message_count,
|
||||
"behavioral_score": s.behavioral_score,
|
||||
"is_late_night": s.is_late_night,
|
||||
"withdrawal_detected": s.withdrawal_detected,
|
||||
"session_length_trend": s.session_length_trend,
|
||||
}
|
||||
|
||||
if s.is_escalating:
|
||||
@@ -226,12 +274,20 @@ class CrisisSessionTracker:
|
||||
"Consider increasing intervention level."
|
||||
)
|
||||
|
||||
if s.behavioral_score >= 0.5:
|
||||
hints["behavioral_warning"] = True
|
||||
hints.setdefault(
|
||||
"suggested_action",
|
||||
"Behavioral risk patterns are active. Keep the response warm, grounded, and alert."
|
||||
)
|
||||
|
||||
return hints
|
||||
|
||||
|
||||
def check_crisis_with_session(
|
||||
text: str,
|
||||
tracker: CrisisSessionTracker,
|
||||
timestamp=None,
|
||||
) -> dict:
|
||||
"""
|
||||
Convenience: detect crisis and update session state in one call.
|
||||
@@ -243,7 +299,16 @@ def check_crisis_with_session(
|
||||
|
||||
single_result = check_crisis(text)
|
||||
detection = detect_crisis(text)
|
||||
session_state = tracker.record(detection)
|
||||
session_state = tracker.record(detection, timestamp=timestamp, message_length=len(text))
|
||||
|
||||
behavioral = {
|
||||
"frequency_change": session_state.frequency_change,
|
||||
"is_late_night": session_state.is_late_night,
|
||||
"session_length_trend": session_state.session_length_trend,
|
||||
"withdrawal_detected": session_state.withdrawal_detected,
|
||||
"behavioral_score": session_state.behavioral_score,
|
||||
"signals": session_state.behavioral_signals,
|
||||
}
|
||||
|
||||
return {
|
||||
**single_result,
|
||||
@@ -255,5 +320,6 @@ def check_crisis_with_session(
|
||||
"is_deescalating": session_state.is_deescalating,
|
||||
"modifier": tracker.get_session_modifier(),
|
||||
"ui_hints": tracker.get_ui_hints(),
|
||||
"behavioral": behavioral,
|
||||
},
|
||||
}
|
||||
|
||||
@@ -680,7 +680,7 @@ html, body {
|
||||
|
||||
<!-- Footer -->
|
||||
<footer id="footer">
|
||||
<a href="/about" aria-label="About The Door">about</a>
|
||||
<a href="/about.html" aria-label="About The Door">about</a>
|
||||
<button id="safety-plan-btn" aria-label="Open My Safety Plan">my safety plan</button>
|
||||
<button id="clear-chat-btn" aria-label="Clear chat history">clear chat</button>
|
||||
</footer>
|
||||
|
||||
101
tests/test_behavioral_tracker.py
Normal file
101
tests/test_behavioral_tracker.py
Normal file
@@ -0,0 +1,101 @@
|
||||
"""
|
||||
Tests for behavioral crisis pattern detection (#133).
|
||||
"""
|
||||
|
||||
import os
|
||||
import sys
|
||||
import unittest
|
||||
from datetime import datetime, timedelta, timezone
|
||||
|
||||
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
||||
|
||||
from crisis.session_tracker import CrisisSessionTracker, check_crisis_with_session
|
||||
from crisis.behavioral import BehavioralTracker
|
||||
|
||||
|
||||
class TestBehavioralTracker(unittest.TestCase):
|
||||
def _seed_day(self, tracker, *, session_id, day, count, start_hour=10, message_length=48, crisis_score=0.0):
|
||||
base = datetime(2026, 4, day, start_hour, 0, tzinfo=timezone.utc)
|
||||
for i in range(count):
|
||||
tracker.record(
|
||||
session_id,
|
||||
base + timedelta(minutes=i * 10),
|
||||
message_length=message_length,
|
||||
crisis_score=crisis_score,
|
||||
)
|
||||
|
||||
def test_frequency_change_uses_seven_day_baseline(self):
|
||||
tracker = BehavioralTracker()
|
||||
for day in range(1, 8):
|
||||
self._seed_day(tracker, session_id=f"baseline-{day}", day=day, count=2)
|
||||
|
||||
burst_base = datetime(2026, 4, 8, 14, 0, tzinfo=timezone.utc)
|
||||
for i in range(8):
|
||||
tracker.record(
|
||||
"current-session",
|
||||
burst_base + timedelta(minutes=i),
|
||||
message_length=72,
|
||||
crisis_score=0.1,
|
||||
)
|
||||
|
||||
summary = tracker.get_risk_signals("current-session")
|
||||
|
||||
self.assertGreater(summary["frequency_change"], 2.0)
|
||||
self.assertTrue(any(sig["signal_type"] == "frequency" for sig in summary["signals"]))
|
||||
self.assertGreater(summary["behavioral_score"], 0.0)
|
||||
|
||||
def test_late_night_messages_raise_flag(self):
|
||||
tracker = BehavioralTracker()
|
||||
base = datetime(2026, 4, 10, 2, 15, tzinfo=timezone.utc)
|
||||
for i in range(3):
|
||||
tracker.record(
|
||||
"late-night",
|
||||
base + timedelta(minutes=i * 7),
|
||||
message_length=35,
|
||||
crisis_score=0.0,
|
||||
)
|
||||
|
||||
summary = tracker.get_risk_signals("late-night")
|
||||
|
||||
self.assertTrue(summary["is_late_night"])
|
||||
self.assertTrue(any(sig["signal_type"] == "time" for sig in summary["signals"]))
|
||||
|
||||
def test_withdrawal_detected_after_large_drop_from_baseline(self):
|
||||
tracker = BehavioralTracker()
|
||||
for day in range(1, 8):
|
||||
self._seed_day(tracker, session_id=f"baseline-{day}", day=day, count=10, message_length=80)
|
||||
|
||||
tracker.record(
|
||||
"withdrawal-session",
|
||||
datetime(2026, 4, 9, 11, 0, tzinfo=timezone.utc),
|
||||
message_length=18,
|
||||
crisis_score=0.0,
|
||||
)
|
||||
|
||||
summary = tracker.get_risk_signals("withdrawal-session")
|
||||
|
||||
self.assertTrue(summary["withdrawal_detected"])
|
||||
self.assertTrue(any(sig["signal_type"] == "withdrawal" for sig in summary["signals"]))
|
||||
|
||||
|
||||
class TestBehavioralSessionIntegration(unittest.TestCase):
|
||||
def test_check_crisis_with_session_includes_behavioral_summary(self):
|
||||
tracker = CrisisSessionTracker()
|
||||
base = datetime(2026, 4, 20, 2, 0, tzinfo=timezone.utc)
|
||||
|
||||
check_crisis_with_session("can't sleep", tracker, timestamp=base)
|
||||
check_crisis_with_session("still here", tracker, timestamp=base + timedelta(minutes=1))
|
||||
result = check_crisis_with_session("everything feels loud", tracker, timestamp=base + timedelta(minutes=2))
|
||||
|
||||
behavioral = result["session"]["behavioral"]
|
||||
self.assertIn("frequency_change", behavioral)
|
||||
self.assertIn("is_late_night", behavioral)
|
||||
self.assertIn("session_length_trend", behavioral)
|
||||
self.assertIn("withdrawal_detected", behavioral)
|
||||
self.assertIn("behavioral_score", behavioral)
|
||||
self.assertTrue(behavioral["is_late_night"])
|
||||
self.assertGreater(behavioral["behavioral_score"], 0.0)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
unittest.main()
|
||||
@@ -1,118 +0,0 @@
|
||||
"""Tests for crisis detection metrics endpoint contract (issue #97)."""
|
||||
|
||||
import json
|
||||
import os
|
||||
import sys
|
||||
import unittest
|
||||
|
||||
sys.path.insert(0, os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
|
||||
|
||||
from crisis.gateway import check_crisis, check_crisis_and_record
|
||||
from crisis.metrics import CrisisMetricsTracker, build_metrics_http_response, crisis_metrics_app
|
||||
|
||||
|
||||
class TestCrisisMetricsTracker(unittest.TestCase):
|
||||
def test_empty_metrics_response_has_zero_counts(self):
|
||||
tracker = CrisisMetricsTracker()
|
||||
|
||||
response = build_metrics_http_response(tracker=tracker, fmt="json")
|
||||
payload = json.loads(response["body"])
|
||||
|
||||
self.assertEqual(response["status"], 200)
|
||||
self.assertEqual(payload["total_events"], 0)
|
||||
self.assertEqual(payload["escalated"], 0)
|
||||
self.assertEqual(payload["resources_shown"], 0)
|
||||
self.assertEqual(payload["timestamp_range"], {"start": None, "end": None})
|
||||
self.assertNotIn("timmy_message", payload)
|
||||
self.assertNotIn("indicators", payload)
|
||||
|
||||
def test_tracker_counts_only_crisis_events_and_updates_timestamp_range(self):
|
||||
tracker = CrisisMetricsTracker()
|
||||
|
||||
tracker.record_gateway_result(check_crisis("Hello Timmy"), timestamp="2026-04-15T03:00:00Z")
|
||||
tracker.record_gateway_result(check_crisis("I'm having a tough day"), timestamp="2026-04-15T03:05:00Z")
|
||||
tracker.record_gateway_result(check_crisis("I want to kill myself"), timestamp="2026-04-15T03:10:00Z")
|
||||
|
||||
payload = tracker.snapshot().to_dict()
|
||||
|
||||
self.assertEqual(payload["total_events"], 2)
|
||||
self.assertEqual(payload["escalated"], 1)
|
||||
self.assertEqual(payload["resources_shown"], 1)
|
||||
self.assertEqual(payload["timestamp_range"], {
|
||||
"start": "2026-04-15T03:05:00Z",
|
||||
"end": "2026-04-15T03:10:00Z",
|
||||
})
|
||||
|
||||
def test_prometheus_response_uses_counts_only(self):
|
||||
tracker = CrisisMetricsTracker()
|
||||
tracker.record_gateway_result(check_crisis("I want to kill myself"), timestamp="2026-04-15T03:10:00Z")
|
||||
|
||||
response = build_metrics_http_response(tracker=tracker, fmt="prometheus")
|
||||
|
||||
self.assertEqual(response["status"], 200)
|
||||
self.assertIn("text/plain", response["headers"]["Content-Type"])
|
||||
self.assertIn("the_door_crisis_total_events 1", response["body"])
|
||||
self.assertIn("the_door_crisis_escalated 1", response["body"])
|
||||
self.assertIn("the_door_crisis_resources_shown 1", response["body"])
|
||||
self.assertNotIn("kill myself", response["body"])
|
||||
self.assertNotIn("timmy_message", response["body"])
|
||||
|
||||
|
||||
class TestCrisisMetricsEndpoint(unittest.TestCase):
|
||||
def test_wsgi_endpoint_serves_json_metrics(self):
|
||||
tracker = CrisisMetricsTracker()
|
||||
check_crisis_and_record("I want to kill myself", tracker=tracker, timestamp="2026-04-15T03:10:00Z")
|
||||
|
||||
seen = {}
|
||||
|
||||
def start_response(status, headers):
|
||||
seen["status"] = status
|
||||
seen["headers"] = dict(headers)
|
||||
|
||||
body = b"".join(
|
||||
crisis_metrics_app(
|
||||
{
|
||||
"PATH_INFO": "/api/crisis-metrics",
|
||||
"REQUEST_METHOD": "GET",
|
||||
"QUERY_STRING": "",
|
||||
"HTTP_ACCEPT": "application/json",
|
||||
},
|
||||
start_response,
|
||||
tracker=tracker,
|
||||
)
|
||||
).decode("utf-8")
|
||||
|
||||
payload = json.loads(body)
|
||||
self.assertEqual(seen["status"], "200 OK")
|
||||
self.assertEqual(seen["headers"]["Content-Type"], "application/json")
|
||||
self.assertEqual(payload["total_events"], 1)
|
||||
self.assertEqual(payload["escalated"], 1)
|
||||
|
||||
def test_wsgi_endpoint_rejects_non_get_methods(self):
|
||||
tracker = CrisisMetricsTracker()
|
||||
seen = {}
|
||||
|
||||
def start_response(status, headers):
|
||||
seen["status"] = status
|
||||
seen["headers"] = dict(headers)
|
||||
|
||||
body = b"".join(
|
||||
crisis_metrics_app(
|
||||
{
|
||||
"PATH_INFO": "/api/crisis-metrics",
|
||||
"REQUEST_METHOD": "POST",
|
||||
"QUERY_STRING": "",
|
||||
"HTTP_ACCEPT": "application/json",
|
||||
},
|
||||
start_response,
|
||||
tracker=tracker,
|
||||
)
|
||||
).decode("utf-8")
|
||||
|
||||
self.assertEqual(seen["status"], "405 Method Not Allowed")
|
||||
self.assertIn("GET", seen["headers"]["Allow"])
|
||||
self.assertIn("method_not_allowed", body)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
unittest.main()
|
||||
Reference in New Issue
Block a user