AI SAFE² (Secure AI Framework for Enterprise Ecosystems) is the open-source governance standard for Security Architects, GRC Officers, and Platform Engineers who need to secure Agentic AI without slowing down innovation.
While legacy GRC platforms (e.g., Vanta, Drata) are still struggling to define what an "AI Agent" is, AI SAFE² v2.1 provides the finished bridge between the Technical Reality of your stack and the Regulatory Reality of the world.
What this Means: AI SAFE² is the only framework that maps Agentic Swarms, Non-Human Identities (NHI), and Vector DB risks directly to ISO 42001, NIST AI RMF, SOC 2 and 10+ other frameworks controls into a single operational schema.
The framework is architected around 5 Strategic Pillars that provide defense-in-depth for Agentic Workflows.
-
The Shield: Input validation, prompt injection defense, and cryptographic agent sandboxing.
-
The Ledger: Full visibility, immutable logging (Chain of Thought), and asset registry.
-
The Brakes: Kill switches, circuit breakers, and "Safe Mode" reversion protocols.
-
The Control Room: Human-in-the-loop (HITL) workflows and real-time anomaly detection.
-
The Feedback Loop: Continuous Red Teaming, threat intelligence integration, and operator training.
Most tools solve one layer of the problem. AI SAFE² is the Unified Operating System that connects them all.
| Competitor Type | Examples | The Gap | The AI SAFE² Solution |
|---|---|---|---|
| Runtime Scanners | Rebuff, PromptMap, Lakera | Too Narrow: They block injections but ignore governance, policy, and compliance evidence. | Governance Layer: We mandate where to put scanners and how to audit them for ISO 42001. |
| Legacy GRC | Vanta, Drata, Archer | Too Broad: They focus on human SSO and laptops. They are blind to Agentic Swarms and RAG Poisoning. | Agentic Native: We treat AI Agents as "First-Class Citizens" with their own identity and access lifecycles. |
| General Frameworks | NIST AI RMF, MITRE ATLAS | Too Abstract: They tell you what the risks are, but not how to engineer the fix in Python/Docker. | Engineering SOPs: We provide the JSON schemas, CLI commands, and configs to enforce the standard. |
AI SAFE² v2.1 is engineered for the Modern AI Stack, not just chatbots. It explicitly governs:
- Orchestration: n8n, LangGraph, Make.com, CrewAI and others.
- Coding Assistants: Cursor, Windsurf ("Vibe Coding"), GitHub Copilot and others.
- Infrastructure: Vector Databases (Pinecone/Weaviate), MCP Servers, and Dockerized Agents.
The 95% Advantage: We have done the heavy lifting. By adopting AI SAFE², you inherit a pre-built governance structure that satisfies ISO 42001, NIST AI RMF, SOC 2, HIPAA, and GDPR simultaneously.
Most frameworks stop at the model. AI SAFE² v2.1 explicitly models and mandates controls across the entire real-world stack, securing the tools your developers actually use (Cursor, Make, LangGraph, AutoGen).
| Layer | Scope | Key Controls |
|---|---|---|
| L1: Core Models | LLMs, Fine-Tuned Weights | [P1.T1.2_ADV] OpenSSF Model Signing |
| L2: Data Infra | Vector DBs, Data Lakes | [P1.T1.5_ADV] Memory Poisoning Defense |
| L3: Patterns | RAG, MCP, API Integrations | [P2.T1.4_ADV] Context Integrity Verification |
| L4: Agentic AI | Swarms, Orchestration (n8n) | [P3.T1.1_ADV] Distributed Kill Switches |
| L5: NHI Users | Service Accounts, Agents | [P1.T2.2_ADV] Non-Human Identity Governance |
AI SAFE² isn't just a list; it's a multi-dimensional defense grid.
Legend: Green = Dedicated Control | 🔗 = Inherited Coverage
graph LR;
A[User Input / Agent Action] -->|Interception| B{Pillar 1: Firewall};
B -- "Injection Detected" --> C[BLOCK & LOG];
B -- "Clean" --> D{Pillar 2: Policy Check};
D -- "Violation" --> C;
D -- "Approved" --> E[Model Inference];
E --> F{Pillar 3: Output Scan};
F -- "Secret Leaked" --> G[Redact & Alert];
F -- "Safe" --> H[Execute Action];
style C fill:#B80000,stroke:#333,stroke-width:2px;
style H fill:#006400,stroke:#333,stroke-width:2px;
The framework is engineered for simultaneous compliance reporting. A single implementation of AI SAFE² controls satisfies the requirements of multiple global standards, eliminating the need for fragmented governance initiatives.
| Standard | Coverage | Why It Matters |
|---|---|---|
| NIST AI RMF | 100% | Maps to all 4 dimensions: Govern, Map, Measure, Manage. |
| ISO/IEC 42001 | 100% | Covers all key clauses (§ 8.1–8.6) and Annex A/B for AI Management Systems. |
| MIT AI Risk Repo | 100% | Integrates all 7 domains and 1,600+ risks (April 2025), including new multi-agent vectors. |
| OWASP LLM | 100% | Complete alignment with LLM01–LLM10 (Prompt Injection, Excessive Agency). |
| MITRE ATLAS | 98% | Covers 10 tactics + 14 new agent-focused techniques (Oct 2025). |
| Google SAIF | 95% | AI SAFE² is superior in Swarm Security, NHI Governance, and Memory Poisoning. |
| CSETv1 Harm | 92% | Maps to all 8 harm types (Physical Safety, Financial Loss, Democratic Norms). |
| SOC 2 Type II | Aligned | Direct mapping for CC.7.1 (Monitoring), CC.6.1 (Access), and A.1.2 (Availability). |
| ISO 27001:2022 | Aligned | Maps to Annex A controls for Access (A.9.2), Logging (A.12.4), and Resilience (A.17.2). |
| NIST CSF | Aligned | Pillars map directly to the Core: Identify (P2), Protect (P1), Detect (P4), Recover (P3). |
| HIPAA | Aligned | Enforces PHI Redaction [P1.T1.5] and Disaster Recovery [P3.T6] (§ 164.308). |
| GDPR | Aligned | Enforces Data Minimization, Storage Limitation, and Lawful Processing. |
| CVE / CVSS | Integrated | Combined Risk Score Formula: CVSS + (100 - Pillar Score) / 10. |
| Zero Trust | Native | Built on "Never Trust, Always Verify" for Non-Human Identities. |
- MIT AI Risk Repository: v2.1 is the first framework to achieve 100% mapping to the repository's 1,600+ cataloged risks.
- CVE/CVSS Integration: Unlike static frameworks, AI SAFE² uses technical vulnerability scores to adjust organizational risk. (e.g., A High CVSS (8.5) combined with Medium Control Effectiveness results in a Critical Risk).
- Google SAIF Superiority: While SAIF provides a baseline, AI SAFE² provides explicit "Gap Filler" sub-domains for Agentic AI and RAG Security that SAIF only implies.
- Foundational Security: We treat ISO 27001 & NIST CSF as the general security foundation, mapping the AI-specific S-A-F-E-E pillars directly into standard enterprise operations.
Why use AI SAFE² vs. trying to patch together legacy tools?
| Feature / Capability | AI SAFE² v2.1 (The OS) | Legacy GRC (Vanta, Archer) |
AI Point Tools (Guardrails) |
|---|---|---|---|
| Universal Mapping | ✅ 1-to-Many One control satisfies ISO, NIST, SOC2, HIPAA, PCI-DSS & GDPR instantly. |
Strong on SOC2, but zero-coverage for AI-specific frameworks (e.g. MITRE ATLAS or MIT AI Risk). You have to manually bridge the gap. |
❌ None Technical blocks only. No compliance evidence. Useless for an auditor. |
| Agentic Awareness | ✅ Native Governs Swarms, n8n loops, and "Vibe Coding" IDEs (Cursor). |
❌ Blind Treats AI as generic software. No concept of autonomous decision loops or multi-agent risks. |
Can monitor LLM I/O, but lacks context on the orchestration or workflow logic (e.g., n8n loops). |
| Non-Human Identity | ✅ First-Class Citizen Dedicated governance for the 100x explosion of Agent API keys. |
❌ Human-Centric Focuses on SSO/MFA for people, Completely misses the 100x explosion of Service Accounts and Agent API keys. |
Scans for secrets, but doesn't govern lifecycle or privilege of the Agent identity. |
| Supply Chain | ✅ Cryptographic Mandates OpenSSF Model Signing & SBOMs. Blocks "Model Pickling" attacks. |
Relies on vendor trust vs. technical verification. |
✅ Strong Good scanning, but lacks policy enforcement. |
| Cognitive Security (Memory) | ✅ Anti-Poisoning Architecture Specific controls for RAG Poisoning (AgentPoison/MINJA). Mandates cryptographic state verification & semantic drift detection. |
❌ Zero Coverage Legacy tools do not understand Vector Databases or Embedding integrity. Let alone why memory systems like RAG, CAG or fine tuning are important. |
Can filter bad inputs, but rarely monitors the long-term memory state of the Vector DB for corruption. |
| Implementation | ✅ 60 Minutes Download the Toolkit. Audit-ready immediately. |
❌ 6-12 Months Requires expensive sales cycles and manual config of custom controls. |
❌ High Friction Requires code integration before value is realized. |
The Verdict: You can keep looking for a tool that catches up to AI SAFE², or you can adopt the standard that defined the race.
This repository contains the definitions (The "What"). To operationalize this in an Enterprise (The "How"), use the Implementation Toolkit.
| Asset | Description | Access |
|---|---|---|
| Taxonomy Definitions | Full Markdown descriptions of controls. | ✅ Free (This Repo) |
| Audit Scorecard | Excel-based calculator with 128 controls & risk formulas. | 🔒 Get Toolkit |
| Governance Policy | MS Word Legal Template mapped to ISO 42001. | 🔒 Get Toolkit |
| Engineering SOPs | CLI commands and configs for Sanitize & Isolate (P1). | 🔒 Get Toolkit |
| Dev-Ready Pack | JSON Schemas & Local MCP Server Scripts. | 🔒 Get Toolkit |
Includes Excel Scorecards, Legal Policies, and the Risk Command Center Dashboard.
AI SAFE² is a living standard that adapts to the threat landscape.
| Version | Focus | Key Metaphor | Control Depth |
|---|---|---|---|
| v2.1 | Agentic & Distributed | Mission Control | 128 Controls (Swarm, NHI, Memory) |
| v2.0 | Enterprise Operations | Building Security | 99 Controls (NIST/ISO Mapping) |
| v1.0 | Foundational Concepts | The Blueprint | 10 Topics (Conceptual) |
👉 Read the Full Evolution History & Changelog
/
├── .github/ # CI/CD Workflows & Dependabot Config
├── assets/ # Visual Maps & Diagrams
├── examples/ # Developer Quickstarts (Python Scripts)
│ └── quickstart_nhi_check.py
├── research/ # Deep Dive Evidence & Justifications
│ ├── 001_rag_poisoning.md
│ ├── 002_nhi_secret_sprawl.md
│ ├── 008_grc_framework_comparison.md
│ └── ...
├── resources/ # Free Community Tools
│ └── pillar1_checklist_lite.md
├── taxonomy/ # The 5 Pillars (Core Framework Definitions)
│ ├── 01_sanitize_isolate.md
│ ├── 02_audit_inventory.md
│ ├── 03_failsafe_recovery.md
│ ├── 04_engage_monitor.md
│ └── 05_evolve_educate.md
├── .markdownlint.yaml # Linter Configuration Rules
├── CITATION.cff # Academic Citation File
├── CODE_OF_CONDUCT.md # Professional Standards
├── CODEOWNERS # Governance Enforcement Rules
├── CONTRIBUTING.md # RFC Process & Proposal Guidelines
├── EVOLUTION.md # Version History (v1.0 -> v2.1)
├── LICENSE # Dual License Text
├── MAINTAINERS.md # Project Leadership
├── README.md # The Universal GRC Standard (You are here)
├── SECURITY.md # Vulnerability Reporting Policy
└── VANGUARD_PROGRAM.md # Community Tiers & Beta Access
This isn't just a repo; it's a mission. We recognize and reward the top 1% of security engineers who contribute to the standard.
- ⭐ Star the Repo: Unlock the "Supporter" role.
- 💡 Contribute: Submit a PR to earn "Contributor" status.
- 🏆 The Vanguard: Earn Priority Beta Access to Agentic Shield (SaaS) by helping us harden the framework.
Read the Vanguard Program Details
If you use AI SAFE² in research or commercial tooling, please cite the Cyber Strategy Institute:
@misc{aisafe2_framework,
title = {AI SAFE² Framework v2.1: The Universal GRC Standard for Agentic AI},
author = {Sullivan, Vincent and {Cyber Strategy Institute}},
year = {2025-2026},
publisher = {Cyber Strategy Institute},
url = {https://github.com/CyberStrategy1/ai-safe2-framework},
note = {Version 2.1. Agentic & Distributed Edition}
}
This project uses a Dual-License Model to support both open innovation and standardized governance.
Applies to: MCP Server scripts, JSON schemas, HTML dashboards, and code snippets. You Can: Use this code commercially, modify it, close-source your modifications, and sell software built with it. The Intent: Build products on top of this. We want this to be the infrastructure of the AI industry.
Applies to: The "AI SAFE²" methodology text, pillar definitions, and PDF manuals. You Can: Share, copy, and redistribute the material. You can adapt it for your internal needs. You Must: Attribution: Give credit to Cyber Strategy Institute. ShareAlike: If you create a public derivative (e.g., "AI SAFE v3.0"), you must share those improvements back to the community under this same license.
Copyright © 2025-2026. All Rights Reserved.
