Mistral Small
Mistral Small
Version: 1
Mistral AILast updated August 2025
Mistral Small can be used on any language-based task that requires high efficiency and low latency.
Low latency
Multilingual

Key capabilities

About this model

Mistral Small is Mistral AI's most efficient Large Language Model (LLM). It can be used on any language-based task that requires high efficiency and low latency.

Key model capabilities

Mistral Small is:
  • A small model optimized for low latency. Very efficient for high volume and low latency workloads. Mistral Small is Mistral's smallest proprietary model, it outperforms Mixtral 8x7B and has lower latency.
  • Specialized in RAG. Crucial information is not lost in the middle of long context windows (up to 32K tokens).
  • Strong in coding. Code generation, review and comments. Supports all mainstream coding languages.
  • Multi-lingual by design. Best-in-class performance in French, German, Spanish, and Italian - in addition to English. Dozens of other languages are supported.
  • Responsible AI. Efficient guardrails baked in the model, with additional safety layer with safe_mode option

Use cases

See Responsible AI for additional considerations for responsible use.

Key use cases

The provider has not supplied this information.

Out of scope use cases

Prompts and completions are passed through a default configuration of Azure AI Content Safety classification models to detect and prevent the output of harmful content. Learn more about Azure AI Content Safety . Configuration options for content filtering vary when you deploy a model for production in Azure AI; learn more .

Pricing

Pricing is based on a number of factors, including deployment type and tokens used. See pricing details here.

Technical specs

The provider has not supplied this information.

Training cut-off date

The provider has not supplied this information.

Training time

The provider has not supplied this information.

Input formats

The provider has not supplied this information.

Output formats

The provider has not supplied this information.

Supported languages

Best-in-class performance in French, German, Spanish, and Italian - in addition to English. Dozens of other languages are supported.

Sample JSON response

The provider has not supplied this information.

Model architecture

The provider has not supplied this information.

Long context

Crucial information is not lost in the middle of long context windows (up to 32K tokens).

Optimizing model performance

The provider has not supplied this information.

Additional assets

For full details of this model, please read release blog post .

Training disclosure

Training, testing and validation

The provider has not supplied this information.

Distribution

Distribution channels

The provider has not supplied this information.

More information

The provider has not supplied this information.

Responsible AI considerations

Safety techniques

Efficient guardrails baked in the model, with additional safety layer with safe_mode option Prompts and completions are passed through a default configuration of Azure AI Content Safety classification models to detect and prevent the output of harmful content. Learn more about Azure AI Content Safety . Configuration options for content filtering vary when you deploy a model for production in Azure AI; learn more .

Safety evaluations

The provider has not supplied this information.

Known limitations

The provider has not supplied this information.

Acceptable use

Acceptable use policy

The provider has not supplied this information.

Quality and performance evaluations

Source: Mistral AI The provider has not supplied this information.

Benchmarking methodology

Source: Mistral AI The provider has not supplied this information.

Public data summary

Source: Mistral AI The provider has not supplied this information.
Model Specifications
Context Length32768
LicenseCustom
Training DataMarch 2023
Last UpdatedAugust 2025
Input TypeText
Output TypeText
ProviderMistral AI
Languages5 Languages