New: Audio API, Embeddings & Realtime WebSocket now available!
osmAPI LogoosmAPI
Integrations

Azure Infrastructure Integration

Connect your Azure OpenAI deployments to the osmAPI orchestration layer for enterprise-grade generative AI.

Azure provides high-availability access to OpenAI's foundation models via Microsoft's enterprise-grade cloud infrastructure. This guide outlines the protocol for provisioning Azure resources, deploying model instances, and industrializing them through the osmAPI gateway.

Technical Scope: Only OpenAI model families are currently supported via the Azure connector. To request support for additional Azure AI services, please submit a capability request.

Integration Prerequisites

  • An active Azure Subscription with OpenAI service permissions.
  • An osmAPI Account with administrative workspace access.

Establishing the Azure Resource

Azure model access is governed by dedicated resource instances and specific deployment identifiers.

Phase 1: Resource Provisioning

Initialize Azure OpenAI Instance

  1. Authenticate with the Azure Portal.
  2. Select Create a Resource and search for Azure OpenAI.
  3. Execute the creation workflow:
    • Subscription: Select your primary enterprise subscription.
    • Region: Define your data residency (e.g., East US, North Europe).
    • Resource Name: Choose a unique identifier (this becomes your <resource-identifier>).
    • Pricing Tier: Standard S0 is the recommended baseline.
  4. Review and deploy the resource.
  5. Note: Your unique resource name is integrated into your API endpoint: https://<resource-name>.openai.azure.com.

Model Deployment & Identifiers

  1. Navigate to your new resource and enter the Azure AI Studio.
  2. Access the Deployments module.
  3. Select Create New Deployment:
    • Model Selection: Choose your target architecture (e.g., gpt-4o, gpt-4-turbo).
    • Deployment Identifier: This name MUST align with the osmAPI model identifier.
    • Deployment Type: AI Foundry is the default deployment type. Global Standard is recommended for cross-boundary scaling.
  4. Finalize the deployment.

Credential Acquisition

  1. Within the Azure Portal, access the Keys and Endpoint section of your resource.
  2. Secure your API Key (Key 1) and your Endpoint URL.

Phase 2: osmAPI Registry

Open the Governance Dashboard

  1. Log into the osmAPI Console.
  2. Select your target Workspace.
  3. Access Provider Governance (Provider Keys) in the system settings.
  1. Select the Add Key action for the Azure module.
  2. Input your Azure API Key.
  3. Input your Resource Identifier (the sub-domain of your Azure endpoint).
  4. Select the Interface Type (Azure OpenAI or AI Foundry). AI Foundry is the default deployment type.
  5. Define a Validation Model from your active deployments to perform an automated handshake verification.
  6. Synchronize the key.

Reliability Verification

Test the new infrastructure segment with a standard orchestration call:

curl -X POST https://api.osmapi.com/v1/chat/completions \
  -H "Authorization: Bearer ${OSM_API_KEY}" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "azure/gpt-4o-mini",
    "messages": [{ "role": "user", "content": "Establish enterprise gateway handshake." }]
  }'

Infrastructure Catalog (Azure)

Once the connection is established, your Azure deployments are exposed via the osmAPI namespace:

  • Enterprise GPT-4o: azure/gpt-4o
  • Efficient GPT-4o Mini: azure/gpt-4o-mini
  • Legacy GPT-3.5 (Deprecated): azure/gpt-3.5-turbo

Alignment Protocol: Only models explicitly deployed within your Azure AI Studio will respond to API requests. Ensure your Deployment Name in Azure perfectly matches the requested model ID.


Operational Forensics & Troubleshooting

Deployment Resolution Failures

  • Verify that the Deployment Name in Azure Studio is a character-perfect match for the osmAPI model identifier.
  • Ensure the resource is in an "Active" state within the Azure Portal.

Regional Availability & Latency

  • Not all model architectures are globally available across all Azure regions. Consult the Azure Model Residency Matrix to optimize your deployment location.
  • For high-availability requirements, consider distributing deployments across multiple Azure regions.

Quota & Throughput Management

  • Azure manages throughput via Tokens Per Minute (TPM). Monitor your consumption in Azure Studio under the Quotas tab.
  • Request enterprise quota expansions directly through Microsoft support to accommodate high-volume osmAPI traffic.

How is this guide?