{"name":"AI Guardrails API","version":"1.0.0","description":"Content safety, PII detection, and hallucination scoring for LLM outputs. Guard your AI pipeline at the edge.","endpoints":{"check":"POST /ai-guardrails/api/check — Full safety check (PII + content safety + hallucination)","pii":"POST /ai-guardrails/api/pii — PII detection only","safety":"POST /ai-guardrails/api/safety — Content safety only","health":"GET /ai-guardrails/health — Health check"},"pii_patterns":5,"safety_keywords":10}