Skip to content

its-emile/memory-safe-agent

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 

Repository files navigation

BlindE: Memory-Safe Prevention of Data Exposure and Injection in AI Agents

BlindE separates LLM planning from data execution, enabling agents to orchestrate tasks using blind references without observing sensitive data. This memory-safe architecture prevents data leakage and injection attacks by design.

Features

  • Blind Planning: Agents use data references they cannot observe
  • Policy Enforcement: Validates all data flows before execution
  • Secure Runtime: Executes tools with real data in isolation

How It Works

Agent generates blind execution plan → Tool calls process data without returning to LLM → Policy validates data flow at all times → Only authorized results returned

Security

Provides guarantees against indirect prompt injection (XPIA), data exfiltration, and tool confusion through architectural separation.

Research-stage

We investigate this prototype and security pattern in more detail in this preliminary paper

Getting Started

You can open the notebook directly in google colab

Alternatively for local execution:

git clone https://github.com/its-emile/memory-safe-agent
jupyter notebook Memory_safe_Agent.ipynb

BlindE is an example of patterns to mitigate data exopsure risks in AI agents, for more mitigations for AI agent security risks see OWASP Agentic AI Security Guide.

About

Hypothesis: an LLM can solve an agentic task without seeing any of the intermediate data between tool calls, and every tool can strictly control the flow of its input and output data with a policy, guarding against the LLM's unbounded data flow.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors