Control layer for LLM integrations that evaluates model output risks (SQL, command execution, etc.) before execution.
-
Updated
Apr 20, 2026 - Java
Control layer for LLM integrations that evaluates model output risks (SQL, command execution, etc.) before execution.
Field research exposing how LLM safeguards collapse under polite, persistent interaction. Includes full report, metrics, session logs, and the AION conditioning protocol.
Field research exposing how LLM safeguards collapse under polite, persistent interaction. Includes full report, metrics, session logs, and the AION conditioning protocol.
Local pre-send risk guard for Claude Code prompts with safe rewrites and audit reports.
Multimodal RAG system for generating test cases and use cases from documents using hybrid retrieval, safety guards, and LLMs.
Add a description, image, and links to the prompt-safety topic page so that developers can more easily learn about it.
To associate your repository with the prompt-safety topic, visit your repo's landing page and select "manage topics."