Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions labs/access-controlling/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Managed Identity
shortDescription: Implement authentication and authorization for AI Gateway APIs.
detailedDescription: Secure your AI Gateway with comprehensive access control mechanisms. This lab covers implementing OAuth 2.0, Azure AD authentication, API keys, and managed identities. Learn to set up subscription-based access, implement JWT validation, configure CORS policies, and create different access tiers for your AI services.
tags: []
authors:
- vieiraae
- nourshaker-msft
Expand Down
1 change: 1 addition & 0 deletions labs/ai-agent-service/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Azure OpenAI
shortDescription: Build and orchestrate AI agents with specialized capabilities.
detailedDescription: Create sophisticated AI agent services that can perform specialized tasks and orchestrate multiple agents. This lab demonstrates how to build agents with specific capabilities (like weather lookup, order placement, or data analysis), use function calling to enable agent actions, and coordinate multiple agents through Azure API Management for complex workflows.
tags: []
authors:
- vieiraae
- nourshaker-msft
Expand Down
1 change: 1 addition & 0 deletions labs/ai-foundry-deepseek/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Deepseek R1
shortDescription: Experiment with Deepseek R1 model via Azure AI Foundry's Model Inference API.
detailedDescription: Playground to try the Deepseek R1 model via the AI Model Inference from Azure AI Foundry. This lab uses the Azure AI Model Inference API and two APIM LLM policies (llm-token-limit and llm-emit-token-metric) to manage and monitor token usage while accessing cutting-edge reasoning models.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/ai-foundry-private-mcp/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,7 @@ services:
- Managed Identity
shortDescription: Create a private network for consuming MCPs using Foundry Agents Classic.
detailedDescription: Playground to show how to create a private network for consuming REST API managed as MCPs using Foundry Agents Classic. This lab demonstrates how to create a private network for consuming MCPs from Foundry Agents Classic using Private Link Services, Azure API Management (APIM), and Azure Front Door.
tags: []
authors:
- pablocast
- nourshaker-msft
Expand Down
1 change: 1 addition & 0 deletions labs/ai-foundry-sdk/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Azure AI Foundry
shortDescription: Use Azure AI Foundry SDK with API Management for enhanced AI applications.
detailedDescription: Learn to leverage the Azure AI Foundry SDK in combination with Azure API Management for building sophisticated AI applications. This lab demonstrates how to use the SDK for model deployment, prompt flow execution, and evaluation, while routing traffic through API Management for governance, security, and monitoring.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/aws-bedrock/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Multi-Cloud AI
shortDescription: Access AWS Bedrock models through Azure API Management gateway.
detailedDescription: Build a multi-cloud AI architecture by integrating AWS Bedrock models with Azure API Management. This lab demonstrates how to expose Claude, Llama, and other Bedrock models through your Azure gateway, implement unified authentication and rate limiting across cloud providers, and create a consistent API interface for multi-cloud AI consumption.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/backend-pool-load-balancing-tf/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ services:
- Azure OpenAI
shortDescription: Built-in load balancing with Terraform - prioritized PTU with fallback consumption scenario.
detailedDescription: Playground to try the built-in load balancing backend pool functionality of APIM to a list of Azure OpenAI endpoints using Terraform for infrastructure as code. This is a typical prioritized PTU with fallback consumption scenario. The lab specifically showcases how a priority 1 (highest) backend is exhausted before gracefully falling back to two equally-weighted priority 2 backends.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/backend-pool-load-balancing/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Azure OpenAI
shortDescription: Built-in load balancing using APIM backend pool functionality across multiple Azure AI Foundry endpoints.
detailedDescription: This lab demonstrates the built-in load balancing backend pool functionality of Azure API Management to distribute traffic across a list of Azure AI Foundry endpoints. This is a typical prioritized PTU with fallback consumption scenario. The lab specifically showcases how a priority 1 (highest) backend is exhausted before gracefully falling back to two equally-weighted priority 2 backends.
tags: []
authors:
- vieiraae
---
Expand Down
3 changes: 2 additions & 1 deletion labs/built-in-logging/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,8 @@ services:
- Application Insights
- Log Analytics
shortDescription: Enable comprehensive logging for debugging and monitoring AI Gateway operations.
detailedDescription: Configure built-in logging capabilities in the AI Gateway to capture detailed request and response data. This lab covers how to set up logging to Application Insights and Log Analytics, configure log levels, implement custom logging policies, and build queries to analyze API traffic patterns and troubleshoot issues.
detailedDescription: Configure built-in logging capabilities in API Management to capture detailed request and response data. This lab covers how to set up logging to Application Insights and Log Analytics, configure log levels, implement custom logging policies, and build queries to analyze API traffic patterns and troubleshoot issues.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/content-safety/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Azure OpenAI
shortDescription: Implement content filtering and safety checks for AI-generated responses.
detailedDescription: This lab demonstrates how to implement Azure AI Content Safety to filter and moderate content in both requests and responses. Learn how to protect your applications from harmful content including hate speech, violence, self-harm, and sexual content using Azure's advanced content moderation capabilities integrated with Azure API Management.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/finops-framework/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,7 @@ services:
- FinOps
shortDescription: Implement comprehensive cost tracking and optimization for AI services.
detailedDescription: Implement a complete FinOps framework for managing AI service costs. This lab covers setting up detailed cost tracking, creating chargeback models for different teams or customers, implementing budget alerts, optimizing token usage, and building dashboards for cost visibility. Learn to balance performance, quality, and cost in your AI deployments.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/function-calling/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Azure OpenAI
shortDescription: Enable AI models to call external functions and APIs for enhanced capabilities.
detailedDescription: Explore the function calling capabilities of Azure OpenAI models integrated through Azure API Management. This lab demonstrates how to define functions, enable AI models to intelligently call external APIs and services, and process the results. Learn to build AI agents that can interact with real-world systems and data sources.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/gemini-mcp-agents/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Multi-Cloud AI
shortDescription: Integrate Google Gemini models with MCP-based agents through Azure API Management.
detailedDescription: Learn to integrate Google Gemini models with Model Context Protocol (MCP) agents through Azure API Management. This lab demonstrates multi-cloud AI architectures, showing how to expose Gemini's capabilities alongside Azure OpenAI, implement MCP servers for tool access, and manage cross-cloud AI services through a unified gateway.
tags: []
authors:
- nourshaker-msft
---
Expand Down
1 change: 1 addition & 0 deletions labs/gemini-models/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Multi-Cloud AI
shortDescription: Access Google Gemini models through Azure API Management with native and OpenAI compatibility.
detailedDescription: Playground to try Gemini Models served through the AI Gateway. This lab creates two APIs - one with native Gemini API compatibility and another one with OpenAI compatibility. Learn to integrate third-party AI models through Azure API Management for unified governance and monitoring.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/google-gemini-api/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Multi-Cloud AI
shortDescription: Access Google Gemini models through OpenAI-compatible endpoint that can be accessed through Azure API Management.
detailedDescription: Playground to try the Google Gemini models with the AI Gateway using the OpenAI-compatible endpoint. This lab demonstrates how to import an OpenAI-compatible Google Gemini API into Azure API Management.
tags: []
authors:
- seilorjunior
---
Expand Down
1 change: 1 addition & 0 deletions labs/image-generation/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Azure OpenAI
shortDescription: Generate images using DALL-E through Azure API Management gateway.
detailedDescription: Learn how to expose and manage Azure OpenAI's DALL-E image generation capabilities through Azure API Management. This lab covers setting up the gateway for image generation requests, implementing size and quality controls, managing costs through rate limiting, and handling image responses efficiently.
tags: []
authors:
- nourshaker-msft
---
Expand Down
1 change: 1 addition & 0 deletions labs/mcp-a2a-agents/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,7 @@ services:
- A2A Protocol
shortDescription: Multi-agent system using Agent-to-Agent protocol with MCP servers and heterogeneous orchestrators.
detailedDescription: Playground to experiment with A2A-enabled agents with Model Context Protocol through Azure API Management. Demonstrates heterogeneous multi-agent systems with agents built using Semantic Kernel and Autogen, communicating through APIM for authentication and authorization. Includes MCP servers for oncall service, weather, and GitHub issues with OAuth integration.
tags: []
authors:
- nourshaker-msft
---
Expand Down
1 change: 1 addition & 0 deletions labs/mcp-client-authorization/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Microsoft Entra ID
shortDescription: Implement MCP client authorization flow with APIM as OAuth client and authorization server.
detailedDescription: Playground to experiment with the Model Context Protocol client authorization flow. In this flow, Azure API Management acts both as an OAuth client connecting to Microsoft Entra ID authorization server and as an OAuth authorization server for the MCP client (MCP inspector). Note: Due to the evolving nature of the MCP Authorization proposal, direct production use is not yet recommended.
tags: []
authors:
- vieiraae
- nourshaker-msft
Expand Down
1 change: 1 addition & 0 deletions labs/mcp-from-api/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ services:
- MCP
shortDescription: Transform existing REST APIs to Model Context Protocol with Azure API Management.
detailedDescription: Playground to transform an existing REST API to the Model Context Protocol with Azure API Management. This lab demonstrates how to expose traditional REST APIs through the MCP standard, enabling AI models to interact with your existing services using the protocol.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/mcp-prm-oauth/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Microsoft Entra ID
shortDescription: Production-grade MCP implementation with OAuth and Protected Resources Metadata (RFC9729).
detailedDescription: Playground to experiment with the Model Context Protocol implementing the MCP Authorization proposal and RFC9729 for Protected Resources Metadata. Azure API Management acts as both OAuth client (to Microsoft Entra ID) and OAuth authorization server (for MCP clients like VS Code or Copilot Studio). This is as close to production-grade security for MCP servers as it gets.
tags: []
authors:
- nourshaker-msft
---
Expand Down
1 change: 1 addition & 0 deletions labs/mcp-registry-apic-github-workflow/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,7 @@ services:
- GitHub Actions
shortDescription: Automated MCP server registry with Azure API Center using GitOps and GitHub CI/CD pipelines.
detailedDescription: Demonstrates a fully automated Model Context Protocol server registry integrating with Azure API Center using GitHub CI/CD pipelines. Developers add MCP servers by committing JSON files, GitHub Actions automatically validate and deploy changes, and Azure API Center becomes the centralized registry for organizational MCP server discovery. Complete GitOps solution with infrastructure as code using Bicep.
tags: []
authors:
- frankqianms
---
Expand Down
1 change: 1 addition & 0 deletions labs/mcp-registry-apic/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- MCP
shortDescription: Centralized MCP server registry using Azure API Center for enterprise governance.
detailedDescription: To unlock the full potential of Model Context Protocol, enterprises need a centralized registry for server discovery and metadata management. Azure API Center serves as a governed, enterprise-grade repository for managing remote MCP servers. This lab demonstrates creating an API Center service and registering example remote MCP servers with centralized oversight for better version control and access management.
tags: []
authors:
- jukasper
---
Expand Down
1 change: 1 addition & 0 deletions labs/message-storing/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Azure Storage
shortDescription: Store conversation history and messages for compliance and analysis.
detailedDescription: Implement message storage solutions to persist conversation history for compliance, auditing, and analysis purposes. This lab shows how to use Azure API Management policies to store messages in Azure Cosmos DB or Azure Storage, manage retention policies, and retrieve conversation history for continuing sessions or compliance reporting.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/model-context-protocol/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Azure OpenAI
shortDescription: Implement the Model Context Protocol for standardized AI integrations.
detailedDescription: Learn to implement the Model Context Protocol (MCP), an open standard for enabling AI models to securely access external data sources and tools. This lab demonstrates how to set up MCP servers, expose them through Azure API Management, and enable AI models to interact with databases, APIs, and other services using this standardized protocol.
tags: []
authors:
- vieiraae
- nourshaker-msft
Expand Down
1 change: 1 addition & 0 deletions labs/model-routing/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Azure OpenAI
shortDescription: Intelligently route requests to different AI models based on content and requirements.
detailedDescription: Implement intelligent model routing strategies to direct requests to the most appropriate AI model based on factors like request content, user tier, cost optimization, or capability requirements. This lab demonstrates how to use Azure API Management policies to analyze incoming requests and route them to different models (GPT-3.5, GPT-4, custom models) for optimal performance and cost efficiency.
tags: []
authors:
- simonkurtz-MSFT
- nourshaker-msft
Expand Down
1 change: 1 addition & 0 deletions labs/openai-agents/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Azure OpenAI
shortDescription: Build AI agents using OpenAI's Assistants API through Azure.
detailedDescription: Explore building AI agents using OpenAI's Assistants API deployed on Azure. This lab demonstrates how to create assistants with custom instructions, enable code interpreter and retrieval capabilities, manage threads and messages, and expose these capabilities through Azure API Management for secure, scalable agent deployments.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/private-connectivity/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- VNet
shortDescription: Secure AI services with private network connectivity using Azure Private Link.
detailedDescription: Implement end-to-end private connectivity for your AI services using Azure Private Link and Virtual Networks. This lab demonstrates how to configure Azure API Management in internal mode, set up private endpoints for Azure OpenAI and other services, implement network security groups, and ensure all traffic remains within your private network for maximum security.
tags: []
authors:
- HoussemDellai
---
Expand Down
1 change: 1 addition & 0 deletions labs/realtime-audio/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Azure OpenAI
shortDescription: Enable real-time audio processing and speech-to-speech AI interactions.
detailedDescription: Build real-time audio applications using Azure OpenAI's audio capabilities through Azure API Management. This lab covers setting up WebSocket connections for streaming audio, implementing speech-to-text and text-to-speech conversions, managing real-time conversations, and handling audio streaming efficiently with low latency.
tags: []
authors:
- nourshaker-msft
---
Expand Down
1 change: 1 addition & 0 deletions labs/realtime-mcp-agents/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,7 @@ services:
- Azure OpenAI
shortDescription: Realtime audio API with MCP tools integration for weather, Spotify, and ServiceNow.
detailedDescription: Playground to experiment with Azure OpenAI Realtime API for text and audio with integrations via Model Context Protocol and Azure API Management. Includes MCP servers for weather service, Spotify (playlists, playback control), and ServiceNow incidents. Leverages the credential manager for OAuth 2.0 tokens and client token validation for end-to-end authentication.
tags: []
authors:
- nourshaker-msft
- vieiraae
Expand Down
1 change: 1 addition & 0 deletions labs/secure-responses-api/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Azure OpenAI
shortDescription: Implement Azure OpenAI Responses API in a secure manner through APIM.
detailedDescription: Playground to try the Azure OpenAI Responses API in a secure manner. This lab demonstrates how to properly secure and expose the Responses API through Azure API Management, ensuring proper authentication, authorization, and monitoring of API responses.
tags: []
authors:
- anotherRedbeard
---
Expand Down
1 change: 1 addition & 0 deletions labs/semantic-caching/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Azure OpenAI
shortDescription: Reduce latency and costs using semantic caching for Azure OpenAI API requests.
detailedDescription: The azure-openai-semantic-cache-lookup policy conducts a cache lookup of responses on Azure OpenAI Chat Completion API and Completion API requests from a pre-configured external cache. It operates by comparing the vector proximity of the prompt to prior requests and using a specific similarity score threshold. Caching responses helps reduce bandwidth and processing demands on the backend Azure OpenAI API, thus reducing latency perceived by API consumers.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/session-awareness/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Azure OpenAI
shortDescription: Maintain conversation context across multiple requests using session management.
detailedDescription: Build session-aware AI applications that maintain conversation context across multiple requests. This lab demonstrates how to implement session management using Azure API Management policies, store session state in Azure Cosmos DB, and provide context-aware responses by including conversation history in subsequent requests to Azure OpenAI.
tags: []
authors:
- dbarkol
---
Expand Down
1 change: 1 addition & 0 deletions labs/slm-self-hosting/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Self-Hosted Gateway
shortDescription: Self-host Phi-3 Small Language Model through APIM self-hosted gateway with OpenAI compatibility.
detailedDescription: Playground to try the self-hosted Phi-3 Small Language Model (SLM) through the APIM self-hosted gateway with OpenAI API compatibility. The Phi-3-Mini-4K-Instruct is a 3.8B parameters, lightweight, state-of-the-art open model. The APIM self-hosted gateway enables use cases where the SLM is running on-premises, bringing the gateway to where your APIs are hosted.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/token-metrics-emitting/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ services:
- Application Insights
shortDescription: Emit and track detailed token usage metrics for cost analysis and optimization.
detailedDescription: Learn how to emit detailed token usage metrics from Azure API Management to Azure Monitor and Application Insights. This lab demonstrates how to track prompt tokens, completion tokens, and total token usage across different models and users. Build dashboards and alerts to monitor costs and usage patterns in real-time.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/token-rate-limiting/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Azure OpenAI
shortDescription: Control costs and prevent abuse with token-based rate limiting policies.
detailedDescription: Implement sophisticated token-based rate limiting to control API usage and costs. This lab shows how to set up rate limits based on token consumption rather than just request counts, providing more granular control over Azure OpenAI usage and costs. Learn to configure different rate limits for different user tiers and implement quota management.
tags: []
authors:
- vieiraae
---
Expand Down
1 change: 1 addition & 0 deletions labs/vector-searching/README.MD
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@ services:
- Azure OpenAI
shortDescription: Implement semantic search using vector embeddings and Azure AI Search.
detailedDescription: Build a semantic search solution using vector embeddings and Azure AI Search. This lab demonstrates how to generate embeddings using Azure OpenAI, store them in Azure AI Search, and implement efficient vector-based search queries through Azure API Management. Learn to build RAG (Retrieval-Augmented Generation) patterns for enhanced AI responses.
tags: []
authors:
- vieiraae
---
Expand Down
Loading
Loading