speakleash-bielik-11b-v3.0-instruct
speakleash-bielik-11b-v3.0-instruct
Version: 1
Hugging FaceLast updated January 2026
Gated Model Access Required speakleash/Bielik-11B-v3.0-Instruct requires special access approval from the authors through Hugging Face. To use this model, you must:
  1. Request access through the model page on Hugging Face and wait for approval from the model authors.
  2. Create a Custom keys workspace connection in Microsoft Foundry or Azure Machine Learning named HuggingFaceTokenConnection with the key HF_TOKEN and value your Hugging Face read or fine-grained token (marked as secret).
  3. Create the Managed Online Endpoint with the property enforce_access_to_default_secret_stores set to enabled so it can access the secret connection value.
  4. Once access is approved, the connection is configured, and the endpoint is created with read access to the token, you can deploy and use the model in Microsoft Foundry or Azure Machine Learning.

speakleash/Bielik-11B-v3.0-Instruct powered by vLLM

Chat Completions API

Send Request

You can use cURL or any REST Client to send a request to the Azure ML endpoint with your Azure ML token.
curl <AZUREML_ENDPOINT_URL> \
    -X POST \
    -H "Authorization: Bearer <AZUREML_TOKEN>" \
    -H "Content-Type: application/json" \
    -d '{"model":"speakleash/Bielik-11B-v3.0-Instruct","messages":[{"role":"user","content":"What is Deep Learning?"}]}'

Supported Parameters

The following are the only mandatory parameters to send in the HTTP POST request to v1/chat/completions.
  • model (string): Model ID used to generate the response, in this case since only a single model is deployed within the same endpoint you can either set it to speakleash/Bielik-11B-v3.0-Instruct or leave it blank instead.
  • messages (array): A list of messages comprising the conversation so far. Depending on the model you use, different message types (modalities) are supported, like text, images, and audio.
The rest of the parameters are optional, and since this model is powered by vLLM with an OpenAI compatible interface on top for the Chat Completions API for text generation, then the I/O interfaces for both generation and streaming are the same as in OpenAI Chat Completions API. You can find the whole specification of the allowed parameters in the OpenAI Chat Completion API Specification , or alternatively in the endpoint /openapi.json for the current Azure ML Endpoint.

Example payload

{
  "model": "speakleash/Bielik-11B-v3.0-Instruct",
  "messages": [
    {"role":"user","content":"What is Deep Learning?"}
  ],
  "max_completion_tokens": 256,
  "temperature": 0.6
}
Model Specifications
LicenseApache-2.0
Last UpdatedJanuary 2026
ProviderHugging Face