Detectors for AI runtime control
Guardion provides multiple detectors to protect your AI systems from various threats and ensure safe, reliable outputs.Available Detectors Types
Prompt Defense
Detect and block prompt injections, jailbreaks, and context hijacking attempts. Backed by ModernGuard models.
Data Protection
Identify and control exposure of personally identifiable information across inputs and outputs.
Content Moderation
Classify and filter unsafe or policy-violating content across multiple safety categories.
Coming Soon
Hallucination Detection
Identify uncertainty in AI responses using representation engineering techniques.
Output Handling
Safely manage code execution and other potentially risky outputs.
More Guardrails
Additional safety features coming to the Guardion platform.
Detector Configuration
Each detector can be configured with custom thresholds and policies through the Guardion dashboard or API.How do detectors work?
How do detectors work?
Guardion’s detectors use advanced machine learning models to analyze inputs and outputs, identifying patterns that match known attack vectors or unsafe content.
Can I customize detection thresholds?
Can I customize detection thresholds?
Yes, each detector allows for custom confidence thresholds to balance security with usability for your specific use case.
How do I implement guardrails?
How do I implement guardrails?
Guardrails can be implemented via our API or SDK. Check our Quickstart guide for implementation details.