Skip to content

Architects of AI SAFE². The Universal GRC Operating System for Agentic AI, Non-Human Identities, and Swarm Governance. [Dual License: MIT + CC-BY-SA]

License

Notifications You must be signed in to change notification settings

CyberStrategyInstitute/ai-safe2-framework

AI SAFE2 Framework Visual Map

AI SAFE² Framework v2.1

The Universal GRC Standard for Agentic AI & ISO 42001 Compliance

Version License: CC BY-SA 4.0 License: MIT Compliance Scope Maintenance

The Narrative | The Architecture | Universal GRC | Comparison | Get Toolkit | Contributing


🛡️ What is AI SAFE²?

AI SAFE² (Secure AI Framework for Enterprise Ecosystems) is the open-source governance standard for Security Architects, GRC Officers, and Platform Engineers who need to secure Agentic AI without slowing down innovation.

🏆 The Race is Over. We Built the Bridge.

While legacy GRC platforms (e.g., Vanta, Drata) are still struggling to define what an "AI Agent" is, AI SAFE² v2.1 provides the finished bridge between the Technical Reality of your stack and the Regulatory Reality of the world.

What this Means: AI SAFE² is the only framework that maps Agentic Swarms, Non-Human Identities (NHI), and Vector DB risks directly to ISO 42001, NIST AI RMF, SOC 2 and 10+ other frameworks controls into a single operational schema.


🏗️ The Core Architecture (The 5 Pillars)

The framework is architected around 5 Strategic Pillars that provide defense-in-depth for Agentic Workflows.

  • P1 The Shield: Input validation, prompt injection defense, and cryptographic agent sandboxing.

  • P2 The Ledger: Full visibility, immutable logging (Chain of Thought), and asset registry.

  • P3 The Brakes: Kill switches, circuit breakers, and "Safe Mode" reversion protocols.

  • P4 The Control Room: Human-in-the-loop (HITL) workflows and real-time anomaly detection.

  • P5 The Feedback Loop: Continuous Red Teaming, threat intelligence integration, and operator training.


🆚 Positioning: Why AI SAFE²? (The "Game Over" Matrix)

Most tools solve one layer of the problem. AI SAFE² is the Unified Operating System that connects them all.

Competitor Type Examples The Gap The AI SAFE² Solution
Runtime Scanners Rebuff, PromptMap, Lakera Too Narrow: They block injections but ignore governance, policy, and compliance evidence. Governance Layer: We mandate where to put scanners and how to audit them for ISO 42001.
Legacy GRC Vanta, Drata, Archer Too Broad: They focus on human SSO and laptops. They are blind to Agentic Swarms and RAG Poisoning. Agentic Native: We treat AI Agents as "First-Class Citizens" with their own identity and access lifecycles.
General Frameworks NIST AI RMF, MITRE ATLAS Too Abstract: They tell you what the risks are, but not how to engineer the fix in Python/Docker. Engineering SOPs: We provide the JSON schemas, CLI commands, and configs to enforce the standard.

🎯 Target Scope & Environment

AI SAFE² v2.1 is engineered for the Modern AI Stack, not just chatbots. It explicitly governs:

  • Orchestration: n8n, LangGraph, Make.com, CrewAI and others.
  • Coding Assistants: Cursor, Windsurf ("Vibe Coding"), GitHub Copilot and others.
  • Infrastructure: Vector Databases (Pinecone/Weaviate), MCP Servers, and Dockerized Agents.

The 95% Advantage: We have done the heavy lifting. By adopting AI SAFE², you inherit a pre-built governance structure that satisfies ISO 42001, NIST AI RMF, SOC 2, HIPAA, and GDPR simultaneously.


🏗️ 5-Layer Architectural Coverage

Most frameworks stop at the model. AI SAFE² v2.1 explicitly models and mandates controls across the entire real-world stack, securing the tools your developers actually use (Cursor, Make, LangGraph, AutoGen).

Layer Scope Key Controls
L1: Core Models LLMs, Fine-Tuned Weights [P1.T1.2_ADV] OpenSSF Model Signing
L2: Data Infra Vector DBs, Data Lakes [P1.T1.5_ADV] Memory Poisoning Defense
L3: Patterns RAG, MCP, API Integrations [P2.T1.4_ADV] Context Integrity Verification
L4: Agentic AI Swarms, Orchestration (n8n) [P3.T1.1_ADV] Distributed Kill Switches
L5: NHI Users Service Accounts, Agents [P1.T2.2_ADV] Non-Human Identity Governance

🏗️ The v2.1 Coverage Matrix

AI SAFE² isn't just a list; it's a multi-dimensional defense grid.

Risk Domain →
Pillar ↓
🤖 Agentic Swarms 🆔 Non-Human Identity 🧠 Memory & RAG 📦 Supply Chain ⚖️ Universal GRC
P1: Sanitize & Isolate P1 P1 P1 P1 P1
P2: Audit & Inventory P2 P2 P2 P2 P2
P3: Fail-Safe & Recovery P3 P3 P3 🔗 P3
P4: ngag & Monitor P4 P4 P4 🔗 P4
P5: Evolve & Educate P5 P5 P5 P5 P5

Legend: Green = Dedicated Control | 🔗 = Inherited Coverage


🧠 The Logic Flow (Infrastructure as Code)

graph LR;
    A[User Input / Agent Action] -->|Interception| B{Pillar 1: Firewall};
    B -- "Injection Detected" --> C[BLOCK & LOG];
    B -- "Clean" --> D{Pillar 2: Policy Check};
    D -- "Violation" --> C;
    D -- "Approved" --> E[Model Inference];
    E --> F{Pillar 3: Output Scan};
    F -- "Secret Leaked" --> G[Redact & Alert];
    F -- "Safe" --> H[Execute Action];
 
style C fill:#B80000,stroke:#333,stroke-width:2px;
style H fill:#006400,stroke:#333,stroke-width:2px;
Loading


🏛️ The "Universal Rosetta Stone"

The framework is engineered for simultaneous compliance reporting. A single implementation of AI SAFE² controls satisfies the requirements of multiple global standards, eliminating the need for fragmented governance initiatives.

Standard Coverage Why It Matters
NIST AI RMF 100% Maps to all 4 dimensions: Govern, Map, Measure, Manage.
ISO/IEC 42001 100% Covers all key clauses (§ 8.1–8.6) and Annex A/B for AI Management Systems.
MIT AI Risk Repo 100% Integrates all 7 domains and 1,600+ risks (April 2025), including new multi-agent vectors.
OWASP LLM 100% Complete alignment with LLM01–LLM10 (Prompt Injection, Excessive Agency).
MITRE ATLAS 98% Covers 10 tactics + 14 new agent-focused techniques (Oct 2025).
Google SAIF 95% AI SAFE² is superior in Swarm Security, NHI Governance, and Memory Poisoning.
CSETv1 Harm 92% Maps to all 8 harm types (Physical Safety, Financial Loss, Democratic Norms).
SOC 2 Type II Aligned Direct mapping for CC.7.1 (Monitoring), CC.6.1 (Access), and A.1.2 (Availability).
ISO 27001:2022 Aligned Maps to Annex A controls for Access (A.9.2), Logging (A.12.4), and Resilience (A.17.2).
NIST CSF Aligned Pillars map directly to the Core: Identify (P2), Protect (P1), Detect (P4), Recover (P3).
HIPAA Aligned Enforces PHI Redaction [P1.T1.5] and Disaster Recovery [P3.T6] (§ 164.308).
GDPR Aligned Enforces Data Minimization, Storage Limitation, and Lawful Processing.
CVE / CVSS Integrated Combined Risk Score Formula: CVSS + (100 - Pillar Score) / 10.
Zero Trust Native Built on "Never Trust, Always Verify" for Non-Human Identities.

🧠 Architectural Insights

  • MIT AI Risk Repository: v2.1 is the first framework to achieve 100% mapping to the repository's 1,600+ cataloged risks.
  • CVE/CVSS Integration: Unlike static frameworks, AI SAFE² uses technical vulnerability scores to adjust organizational risk. (e.g., A High CVSS (8.5) combined with Medium Control Effectiveness results in a Critical Risk).
  • Google SAIF Superiority: While SAIF provides a baseline, AI SAFE² provides explicit "Gap Filler" sub-domains for Agentic AI and RAG Security that SAIF only implies.
  • Foundational Security: We treat ISO 27001 & NIST CSF as the general security foundation, mapping the AI-specific S-A-F-E-E pillars directly into standard enterprise operations.

🆚 Why The Race Is Over (Comparison Matrix)

Why use AI SAFE² vs. trying to patch together legacy tools?

Feature / Capability AI SAFE² v2.1 (The OS) Legacy GRC
(Vanta, Archer)
AI Point Tools
(Guardrails)
Universal Mapping ✅ 1-to-Many
One control satisfies ISO, NIST, SOC2, HIPAA, PCI-DSS & GDPR instantly.
⚠️ Fragmented
Strong on SOC2, but zero-coverage for AI-specific frameworks (e.g. MITRE ATLAS or MIT AI Risk). You have to manually bridge the gap.
❌ None
Technical blocks only. No compliance evidence. Useless for an auditor.
Agentic Awareness ✅ Native
Governs Swarms, n8n loops, and "Vibe Coding" IDEs (Cursor).
❌ Blind
Treats AI as generic software. No concept of autonomous decision loops or multi-agent risks.
⚠️ Partial
Can monitor LLM I/O, but lacks context on the orchestration or workflow logic (e.g., n8n loops).
Non-Human Identity ✅ First-Class Citizen
Dedicated governance for the 100x explosion of Agent API keys.
❌ Human-Centric
Focuses on SSO/MFA for people, Completely misses the 100x explosion of Service Accounts and Agent API keys.
⚠️ Weak
Scans for secrets, but doesn't govern lifecycle or privilege of the Agent identity.
Supply Chain ✅ Cryptographic
Mandates OpenSSF Model Signing & SBOMs. Blocks "Model Pickling" attacks.
⚠️ Questionnaire
Relies on vendor trust vs. technical verification.
✅ Strong
Good scanning, but lacks policy enforcement.
Cognitive Security (Memory) ✅ Anti-Poisoning Architecture
Specific controls for RAG Poisoning (AgentPoison/MINJA). Mandates cryptographic state verification & semantic drift detection.
❌ Zero Coverage
Legacy tools do not understand Vector Databases or Embedding integrity. Let alone why memory systems like RAG, CAG or fine tuning are important.
⚠️ Reactive
Can filter bad inputs, but rarely monitors the long-term memory state of the Vector DB for corruption.
Implementation ✅ 60 Minutes
Download the Toolkit. Audit-ready immediately.
❌ 6-12 Months
Requires expensive sales cycles and manual config of custom controls.
❌ High Friction
Requires code integration before value is realized.

The Verdict: You can keep looking for a tool that catches up to AI SAFE², or you can adopt the standard that defined the race.


🚀 Fast-Track Implementation (The Toolkit)

This repository contains the definitions (The "What"). To operationalize this in an Enterprise (The "How"), use the Implementation Toolkit.

Asset Description Access
Taxonomy Definitions Full Markdown descriptions of controls. Free (This Repo)
Audit Scorecard Excel-based calculator with 128 controls & risk formulas. 🔒 Get Toolkit
Governance Policy MS Word Legal Template mapped to ISO 42001. 🔒 Get Toolkit
Engineering SOPs CLI commands and configs for Sanitize & Isolate (P1). 🔒 Get Toolkit
Dev-Ready Pack JSON Schemas & Local MCP Server Scripts. 🔒 Get Toolkit

Download Toolkit

Includes Excel Scorecards, Legal Policies, and the Risk Command Center Dashboard.


📈 Framework Evolution

AI SAFE² is a living standard that adapts to the threat landscape.

Version Focus Key Metaphor Control Depth
v2.1 Agentic & Distributed Mission Control 128 Controls (Swarm, NHI, Memory)
v2.0 Enterprise Operations Building Security 99 Controls (NIST/ISO Mapping)
v1.0 Foundational Concepts The Blueprint 10 Topics (Conceptual)

👉 Read the Full Evolution History & Changelog

📂 Repository Structure

/
├── .github/                # CI/CD Workflows & Dependabot Config
├── assets/                 # Visual Maps & Diagrams
├── examples/               # Developer Quickstarts (Python Scripts)
│   └── quickstart_nhi_check.py
├── research/               # Deep Dive Evidence & Justifications
│   ├── 001_rag_poisoning.md
│   ├── 002_nhi_secret_sprawl.md
│   ├── 008_grc_framework_comparison.md
│   └── ...
├── resources/              # Free Community Tools
│   └── pillar1_checklist_lite.md
├── taxonomy/               # The 5 Pillars (Core Framework Definitions)
│   ├── 01_sanitize_isolate.md
│   ├── 02_audit_inventory.md
│   ├── 03_failsafe_recovery.md
│   ├── 04_engage_monitor.md
│   └── 05_evolve_educate.md
├── .markdownlint.yaml      # Linter Configuration Rules
├── CITATION.cff            # Academic Citation File
├── CODE_OF_CONDUCT.md      # Professional Standards
├── CODEOWNERS              # Governance Enforcement Rules
├── CONTRIBUTING.md         # RFC Process & Proposal Guidelines
├── EVOLUTION.md            # Version History (v1.0 -> v2.1)
├── LICENSE                 # Dual License Text
├── MAINTAINERS.md          # Project Leadership
├── README.md               # The Universal GRC Standard (You are here)
├── SECURITY.md             # Vulnerability Reporting Policy
└── VANGUARD_PROGRAM.md     # Community Tiers & Beta Access

🤝 Join the Vanguard (Community)

This isn't just a repo; it's a mission. We recognize and reward the top 1% of security engineers who contribute to the standard.

  • ⭐ Star the Repo: Unlock the "Supporter" role.
  • 💡 Contribute: Submit a PR to earn "Contributor" status.
  • 🏆 The Vanguard: Earn Priority Beta Access to Agentic Shield (SaaS) by helping us harden the framework.

Read the Vanguard Program Details


✏️ Citation

If you use AI SAFE² in research or commercial tooling, please cite the Cyber Strategy Institute:

@misc{aisafe2_framework,
  title = {AI SAFE² Framework v2.1: The Universal GRC Standard for Agentic AI},
  author = {Sullivan, Vincent and {Cyber Strategy Institute}},
  year = {2025-2026},
  publisher = {Cyber Strategy Institute},
  url = {https://github.com/CyberStrategy1/ai-safe2-framework},
  note = {Version 2.1. Agentic & Distributed Edition}
}

⚖️ Licensing & Usage Rights

This project uses a Dual-License Model to support both open innovation and standardized governance.

💻 A. The Code: MIT License

Applies to: MCP Server scripts, JSON schemas, HTML dashboards, and code snippets. You Can: Use this code commercially, modify it, close-source your modifications, and sell software built with it. The Intent: Build products on top of this. We want this to be the infrastructure of the AI industry.

📘 B. The Framework/Docs: CC-BY-SA 4.0

Applies to: The "AI SAFE²" methodology text, pillar definitions, and PDF manuals. You Can: Share, copy, and redistribute the material. You can adapt it for your internal needs. You Must: Attribution: Give credit to Cyber Strategy Institute. ShareAlike: If you create a public derivative (e.g., "AI SAFE v3.0"), you must share those improvements back to the community under this same license.

Managed by Cyber Strategy Institute.
Copyright © 2025-2026. All Rights Reserved.

About

Architects of AI SAFE². The Universal GRC Operating System for Agentic AI, Non-Human Identities, and Swarm Governance. [Dual License: MIT + CC-BY-SA]

Topics

Resources

License

Code of conduct

Contributing

Security policy

Stars

Watchers

Forks

Packages

No packages published

Contributors 2

  •  
  •