Azure Infrastructure Integration
Connect your Azure OpenAI deployments to the osmAPI orchestration layer for enterprise-grade generative AI.
Azure provides high-availability access to OpenAI's foundation models via Microsoft's enterprise-grade cloud infrastructure. This guide outlines the protocol for provisioning Azure resources, deploying model instances, and industrializing them through the osmAPI gateway.
Technical Scope: Only OpenAI model families are currently supported via the Azure connector. To request support for additional Azure AI services, please submit a capability request.
Integration Prerequisites
- An active Azure Subscription with OpenAI service permissions.
- An osmAPI Account with administrative workspace access.
Establishing the Azure Resource
Azure model access is governed by dedicated resource instances and specific deployment identifiers.
Phase 1: Resource Provisioning
Initialize Azure OpenAI Instance
- Authenticate with the Azure Portal.
- Select Create a Resource and search for Azure OpenAI.
- Execute the creation workflow:
- Subscription: Select your primary enterprise subscription.
- Region: Define your data residency (e.g., East US, North Europe).
- Resource Name: Choose a unique identifier (this becomes your
<resource-identifier>). - Pricing Tier: Standard S0 is the recommended baseline.
- Review and deploy the resource.
- Note: Your unique resource name is integrated into your API endpoint:
https://<resource-name>.openai.azure.com.
Model Deployment & Identifiers
- Navigate to your new resource and enter the Azure AI Studio.
- Access the Deployments module.
- Select Create New Deployment:
- Model Selection: Choose your target architecture (e.g., gpt-4o, gpt-4-turbo).
- Deployment Identifier: This name MUST align with the osmAPI model identifier.
- Deployment Type: AI Foundry is the default deployment type. Global Standard is recommended for cross-boundary scaling.
- Finalize the deployment.
Credential Acquisition
- Within the Azure Portal, access the Keys and Endpoint section of your resource.
- Secure your API Key (Key 1) and your Endpoint URL.
Phase 2: osmAPI Registry
Open the Governance Dashboard
- Log into the osmAPI Console.
- Select your target Workspace.
- Access Provider Governance (Provider Keys) in the system settings.
Link Azure Credentials
- Select the Add Key action for the Azure module.
- Input your Azure API Key.
- Input your Resource Identifier (the sub-domain of your Azure endpoint).
- Select the Interface Type (Azure OpenAI or AI Foundry). AI Foundry is the default deployment type.
- Define a Validation Model from your active deployments to perform an automated handshake verification.
- Synchronize the key.
Reliability Verification
Test the new infrastructure segment with a standard orchestration call:
curl -X POST https://api.osmapi.com/v1/chat/completions \
-H "Authorization: Bearer ${OSM_API_KEY}" \
-H "Content-Type: application/json" \
-d '{
"model": "azure/gpt-4o-mini",
"messages": [{ "role": "user", "content": "Establish enterprise gateway handshake." }]
}'Infrastructure Catalog (Azure)
Once the connection is established, your Azure deployments are exposed via the osmAPI namespace:
- Enterprise GPT-4o:
azure/gpt-4o - Efficient GPT-4o Mini:
azure/gpt-4o-mini - Legacy GPT-3.5 (Deprecated):
azure/gpt-3.5-turbo
Alignment Protocol: Only models explicitly deployed within your Azure AI
Studio will respond to API requests. Ensure your Deployment Name in Azure
perfectly matches the requested model ID.
Operational Forensics & Troubleshooting
Deployment Resolution Failures
- Verify that the
Deployment Namein Azure Studio is a character-perfect match for the osmAPI model identifier. - Ensure the resource is in an "Active" state within the Azure Portal.
Regional Availability & Latency
- Not all model architectures are globally available across all Azure regions. Consult the Azure Model Residency Matrix to optimize your deployment location.
- For high-availability requirements, consider distributing deployments across multiple Azure regions.
Quota & Throughput Management
- Azure manages throughput via Tokens Per Minute (TPM). Monitor your consumption in Azure Studio under the Quotas tab.
- Request enterprise quota expansions directly through Microsoft support to accommodate high-volume osmAPI traffic.
How is this guide?