granite-4.1-30b
Version: 1
ibm-granite/granite-4.1-30b powered by vLLM
Chat Completions API
Send Request
You can use cURL or any REST Client to send a request to the Azure ML endpoint with your Azure ML token.curl <AZUREML_ENDPOINT_URL> \
-X POST \
-H "Authorization: Bearer <AZUREML_TOKEN>" \
-H "Content-Type: application/json" \
-d '{"model":"ibm-granite/granite-4.1-30b","messages":[{"role":"user","content":"What is Deep Learning?"}]}'
Supported Parameters
The following are the only mandatory parameters to send in the HTTP POST request tov1/chat/completions.
- model (string): Model ID used to generate the response, in this case since only a single model is deployed within the same endpoint you can either set it to ibm-granite/granite-4.1-30b or leave it blank instead.
- messages (array): A list of messages comprising the conversation so far. Depending on the model you use, different message types (modalities) are supported, like text, images, and audio.
/openapi.json for the current Azure ML Endpoint.
Example payload
{
"model": "ibm-granite/granite-4.1-30b",
"messages": [
{"role":"user","content":"What is Deep Learning?"}
],
"max_completion_tokens": 256,
"temperature": 0.6
}
Hugging Face on Foundry
This model is sourced from Hugging Face, which offers thousands of open models for easy deployment on Microsoft Foundry. This model is a Non-Microsoft Product that has not been tested or evaluated by Microsoft. Customers should ensure that the model is appropriate for their specific use, including by evaluating any legal or export-control considerations and conducting their own model risk and safety evaluations. You can learn about Foundry risk and safety evaluations here . You can learn about Hugging Face security measures and requirements for models offered in Foundry here .Model Specifications
LicenseApache-2.0
Last UpdatedApril 2026
ProviderHugging Face