⚠️ Proof of Concept Project: This repository contains a proof-of-concept implementation of an AI-powered infrastructure management agent. It is currently in active development and not intended for production use. We plan to release a production-ready version in the future. Use at your own risk and always test in development environments first.
AI Infrastructure Agent is an intelligent system that allows you to manage AWS infrastructure using natural language commands. Powered by advanced AI models (OpenAI GPT, Google Gemini, or Anthropic Claude), it translates your infrastructure requests into executable AWS operations while maintaining safety through conflict detection and resolution.
- Natural Language Interface - Describe what you want, not how to build it
- Multi-AI Provider Support - Choose between OpenAI, Google Gemini, Anthropic, or AWS Bedrock Nova
- Web Dashboard - Visual interface for infrastructure management, built-in conflict detection and dry-run mode
- Terraform-like state - Maintains accurate infrastructure state
- Current Resource Support - VPC, EC2, SG, Autoscaling Group, ALB. Check the roadmap here: Core Platform Development
Imagine you want to create AWS infrastructure with a simple request:
"Create an EC2 instance for hosting an Apache Server with a dedicated security group that allows inbound HTTP (port 80) and SSH (port 22) traffic."
💡 Amazon Nova Users: When using AWS Bedrock Nova models, you may want to specify the region in your request for better context, e.g., "Create an EC2 instance in us-east-1 for hosting an Apache Server..."
Here's what happens:
The AI agent analyzes your request and creates a detailed execution plan:
sequenceDiagram
participant U as User
participant A as AI Agent
participant S as State Manager
participant M as MCP Server
participant AWS as AWS APIs
U->>A: "Create EC2 instance for Apache Server..."
A->>S: Get current infrastructure state
S->>A: Return current state
A->>M: Query available tools & capabilities
M->>A: Return tool capabilities
A->>A: Generate execution plan with LLM
A->>AWS: Validate plan (dry-run checks)
AWS->>A: Validation results
A->>U: Present execution plan for approval
Note over A,U: Plan includes:<br/>• Get Default VPC<br/>• Create Security Group<br/>• Add HTTP & SSH rules<br/>• Get Latest AMI<br/>• Create EC2 Instance
The agent presents the plan for your review:
- Shows exactly what will be created
- Waits for your approval
Once approved, the agent:
- Creates resources in the correct order
- Monitors progress in real-time
- Handles dependencies automatically
- Reports completion status
- Quick Tutorial: AI Infrastructure Agent for AWS
- Series Tutorial: Building Your Business on AWS with AI Agent
Detailed Guides: Installation Guide
git clone https://github.com/VersusControl/ai-infrastructure-agent.git
cd ai-infrastructure-agent
# Edit the main configuration
nano config.yaml
Choose your preferred AI provider in config.yaml
:
agent:
provider: "openai" # Options: openai, gemini, anthropic, bedrock
model: "gpt-4" # Model to use
max_tokens: 4000
temperature: 0.1
dry_run: true # Start with dry-run enabled
auto_resolve_conflicts: false
Detailed Setup Guides:
- OpenAI: OpenAI API Key Setup Guide
- Google Gemini: Gemini API Key Setup Guide
- AWS Bedrock Nova: AWS Bedrock Nova Configuration Guide
# For OpenAI
export OPENAI_API_KEY="your-openai-api-key"
# For Google Gemini
export GEMINI_API_KEY="your-gemini-api-key"
# For AWS Bedrock Nova - use AWS credentials (no API key needed)
# Configure AWS credentials using: aws configure, environment variables, or IAM roles
# Configure AWS CLI
aws configure
# Or set environment variables
export AWS_ACCESS_KEY_ID="your-access-key"
export AWS_SECRET_ACCESS_KEY="your-secret-key"
export AWS_DEFAULT_REGION="us-west-2"
Basic Docker Run:
docker run -d \
--name ai-infrastructure-agent \
-p 8080:8080 \
-v $(pwd)/config.yaml:/app/config.yaml:ro \
-v $(pwd)/states:/app/states \
-e OPENAI_API_KEY="your-openai-api-key-here" \
-e AWS_ACCESS_KEY_ID="your-aws-access-key" \
-e AWS_SECRET_ACCESS_KEY="your-aws-secret-key" \
-e AWS_DEFAULT_REGION="us-west-2" \
ghcr.io/versuscontrol/ai-infrastructure-agent
Docker Compose (Recommended). Create a docker-compose.yml
file:
version: '3.8'
services:
ai-infrastructure-agent:
image: ghcr.io/versuscontrol/ai-infrastructure-agent
container_name: ai-infrastructure-agent
restart: unless-stopped
ports:
- "8080:8080"
volumes:
# Mount configuration file (read-only)
- ./config.yaml:/app/config.yaml:ro
# Mount data directories (persistent)
- ./states:/app/states
environment:
# AI Provider API Keys (choose one)
- OPENAI_API_KEY=${OPENAI_API_KEY}
# - GEMINI_API_KEY=${GEMINI_API_KEY}
# - ANTHROPIC_API_KEY=${ANTHROPIC_API_KEY}
# AWS Configuration
- AWS_ACCESS_KEY_ID=${AWS_ACCESS_KEY_ID}
- AWS_SECRET_ACCESS_KEY=${AWS_SECRET_ACCESS_KEY}
- AWS_DEFAULT_REGION=${AWS_DEFAULT_REGION:-us-west-2}
Start the application:
# Start with Docker Compose
docker-compose up -d
# View logs
docker-compose logs -f
# Stop the application
docker-compose down
# Clone the repository
git clone https://github.com/VersusControl/ai-infrastructure-agent.git
cd ai-infrastructure-agent
# Run the installation script
./scripts/install.sh
Start the Web UI:
./scripts/run-web-ui.sh
Open your browser and navigate to:
http://localhost:8080
# Simple EC2 instance
"Create a t3.micro EC2 instance with Ubuntu 22.04"
# Web server setup
"Deploy a load-balanced web application with 2 EC2 instances behind an ALB"
# Database setup
"Create an RDS MySQL database with read replicas in multiple AZs"
# Complete environment
"Set up a development environment with VPC, subnets, EC2, and RDS"
Read detail: Technical Architecture Overview
- Web Interface: React-based dashboard for visual interaction
- MCP Server: Core agent implementing Model Context Protocol
- Agent Core: AI-powered decision making and planning
- AWS Client: Secure AWS SDK integration
- State Management: Infrastructure state tracking and conflict resolution
All operations can be run in "dry-run" mode first:
- Shows exactly what would be created/modified/deleted
- Estimates costs before execution
- No actual AWS resources are touched
- Maintains accurate infrastructure state
- Detects drift from expected configuration
- Fork the repository
- Create a feature branch:
git checkout -b feature-name
- Make your changes
- Run tests
- Commit:
git commit -m "Add feature"
- Push:
git push origin feature-name
- Create a Pull Request
AWS Authentication Issues
# Check AWS credentials
aws sts get-caller-identity
# Verify permissions
aws iam get-user
# Test basic AWS access
aws ec2 describe-regions
AI Provider API Issues
# Check API key is set
echo $OPENAI_API_KEY
# Test API connection
curl -H "Authorization: Bearer $OPENAI_API_KEY" \
https://api.openai.com/v1/models
Port Already in Use
# Check what's using the port
lsof -i :8080
lsof -i :3000
# Kill processes if needed
kill -9 <pid>
# Or change ports in config.yaml
Go Build Issues
# Clean module cache
go clean -modcache
# Re-download dependencies
go mod download
go mod tidy
# Rebuild
go build ./...
Decision validation failed: decision confidence too low: 0.000000
Try increase max_tokens:
agent:
provider: "gemini" # Use Google AI (Gemini)
model: "gemini-2.5-flash-lite"
max_tokens: 10000 # <-- increase
- API Keys: Never commit API keys to version control
- AWS Permissions: Use least-privilege IAM policies
- Network Security: Run in private networks when possible
- Audit Logging: Enable comprehensive logging for compliance
- Dry Run: Always test in dry-run mode first
- ✅ Basic natural language processing
- ✅ Core AWS resource management
- ✅ Web dashboard
- ✅ MCP protocol support
- ✅ ReAct-Style Agent
- 🔄 Better UX/UI
- 🔄 Cost optimization recommendations
- 🔄 Enhanced conflict resolution
- 🔄 Infrastructure templates
- 🔄 Multi States
- 🔄 Role-based access control
- GitHub Issues: Report bugs and request features
- Discussions: Community discussions
This project is licensed under the MIT License - see the LICENSE file for details.
This is a proof-of-concept project. While we've implemented safety measures like dry-run mode and conflict detection, always:
- Test in development environments first
- Review all generated plans before execution
- Maintain proper AWS IAM permissions
- Monitor costs and resource usage
- Keep backups of critical infrastructure
The authors are not responsible for any costs, data loss, or security issues that may arise from using this software.
Built with ❤️ by the DevOps VN Team
Empowering infrastructure management through AI