Azure OpenAI Service integration. Manage Models, Deployments, Prompts, Completions. Use when the user wants to interact with Azure OpenAI Service data.
Azure OpenAI Service provides access to OpenAI's powerful language models, including GPT-3, Codex, and DALL-E, through the Azure cloud platform. Developers and organizations use it to build AI-powered applications for natural language processing, code generation, and image creation. It's suitable for businesses seeking enterprise-grade security, compliance, and scalability.
Official docs: https://learn.microsoft.com/en-us/azure/cognitive-services/openai/
Use action names and parameters as needed.
This skill uses the Membrane CLI to interact with Azure OpenAI Service. Membrane handles authentication and credentials refresh automatically — so you can focus on the integration logic rather than auth plumbing.
Install the Membrane CLI so you can run membrane from the terminal:
npm install -g @membranehq/cli
membrane login --tenant
A browser window opens for authentication.
Headless environments: Run the command, copy the printed URL for the user to open in a browser, then complete with membrane login complete <code>.
membrane search azure-openai-service --elementType=connector --json
Take the connector ID from output.items[0].element?.id, then:
membrane connect --connectorId=CONNECTOR_ID --json
The user completes authentication in the browser. The output contains the new connection id.When you are not sure if connection already exists:
membrane connection list --json
If a Azure OpenAI Service connection exists, note its connectionIdWhen you know what you want to do but not the exact action ID:
membrane action list --intent=QUERY --connectionId=CONNECTION_ID --json
This will return action objects with id and inputSchema in it, so you will know how to run it.
| Name | Key | Description |
|---|---|---|
| Create Completion | create-completion | Creates a text completion for the provided prompt using Azure OpenAI. |
| Create Audio Translation | create-audio-translation | Translates audio from any language into English text using Azure OpenAI Whisper models. |
| Create Audio Transcription | create-audio-transcription | Transcribes audio into text using Azure OpenAI Whisper models. |
| Generate Image | generate-image | Generates an image using DALL-E models deployed on Azure OpenAI. |
| Create Embedding | create-embedding | Creates an embedding vector representing the input text. |
| Create Chat Completion | create-chat-completion | Creates a chat completion using the Azure OpenAI API. |
membrane action run --connectionId=CONNECTION_ID ACTION_ID --json
To pass JSON parameters:
membrane action run --connectionId=CONNECTION_ID ACTION_ID --json --input "{ \"key\": \"value\" }"
When the available actions don't cover your use case, you can send requests directly to the Azure OpenAI Service API through Membrane's proxy. Membrane automatically appends the base URL to the path you provide and injects the correct authentication headers — including transparent credential refresh if they expire.
membrane request CONNECTION_ID /path/to/endpoint
Common options:
| Flag | Description |
|---|---|
-X, --method | HTTP method (GET, POST, PUT, PATCH, DELETE). Defaults to GET |
-H, --header | Add a request header (repeatable), e.g. -H "Accept: application/json" |
-d, --data | Request body (string) |
--json | Shorthand to send a JSON body and set Content-Type: application/json |
--rawData | Send the body as-is without any processing |
--query | Query-string parameter (repeatable), e.g. --query "limit=10" |
--pathParam | Path parameter (repeatable), e.g. --pathParam "id=123" |
membrane action list --intent=QUERY (replace QUERY with your intent) to find existing actions before writing custom API calls. Pre-built actions handle pagination, field mapping, and edge cases that raw API calls miss.ZIP package — ready to use