Ministral 3B
Ministral 3B
Version: 1
Mistral AILast updated October 2024
Ministral 3B is a state-of-the-art Small Language Model (SLM) optimized for edge computing and on-device applications. As it is designed for low-latency and compute-efficient inference, it it also the perfect model for standard GenAI applications that have
Low latency
Agents
Reasoning
Ministral 3B is a state-of-the-art Small Language Model (SLM) optimized for edge computing and on-device applications. As it is designed for low-latency and compute-efficient inference, it it also the perfect model for standard GenAI applications that have real-time requirements and high-volume. Number of Parameters: 3,6 billions Ministral 3B and Ministral 8B set a new frontier in knowledge, commonsense, reasoning, function-calling, and efficiency in the sub-10B category, and can be used or tuned to a variety of uses, from orchestrating agentic workflows to creating specialist task workers. Both models support up to 128k context length (currently 32k on vLLM) and Ministral 8B has a special interleaved sliding-window attention pattern for faster and memory-efficient inference.

Use cases

Our most innovative customers and partners have increasingly been asking for local, privacy-first inference for critical applications such as on-device translation, internet-less smart assistants, local analytics, and autonomous robotics. Les Ministraux were built to provide a compute-efficient and low-latency solution for these scenarios. From independent hobbyists to global manufacturing teams, les Ministraux deliver for a wide variety of use cases. Used in conjunction with larger language models such as Mistral Large, les Ministraux are also efficient intermediaries for function-calling in multi-step agentic workflows. They can be tuned to handle input parsing, task routing, and calling APIs based on user intent across multiple contexts at extremely low latency and cost. Source: Un Ministral, des Ministraux - Introducing the world’s best edge models.

Content Filtering

Prompts and completions are passed through a default configuration of Azure AI Content Safety classification models to detect and prevent the output of harmful content. Learn more about Azure AI Content Safety . Configuration options for content filtering vary when you deploy a model for production in Azure AI; learn more .
Source: Un Ministral, des Ministraux - Introducing the world’s best edge models. We demonstrate the performance of les Ministraux across multiple tasks where they consistently outperform their peers. We re-evaluated all models with our internal framework for fair comparison.

Pretrained Models

Knowledge & Commonsense

ModelMMLUAGIEvalWinograndeArc-cTriviaQA
Gemma 2 2B52.433.868.742.647.8
Llama 3.2 3B56.237.459.643.150.7
Ministral 3B60.942.172.764.256.7
Mistral 7B62.442.574.267.962.5
Llama 3.1 8B64.744.474.646.060.2
Ministral 8B65.048.375.371.965.5

Code and Math

ModelHumanEval (pass@1)GSM8K (maj@8)
Gemma 2 2B20.135.5
Llama 3.2 3B29.937.2
Ministral 3B34.250.9
Mistral 7B26.851.3
Llama 3.1 8B37.861.7
Ministral 8B34.864.5

Multilingual

ModelFrench MMLUGerman MMLUSpanish MMLU
Gemma 2 2B41.040.141.7
Llama 3.2 3B42.342.243.1
Ministral 3B49.148.349.5
Mistral 7B50.649.651.4
Llama 3.1 8B50.852.854.6
Ministral 8B57.557.459.6

Instruct Models

Chat/Arena (gpt-4o judge)

ModelMTBenchArena HardWild bench
Gemma 2 2B7.551.732.5
Llama 3.2 3B7.246.027.2
Ministral 3B8.164.336.3
Mistral 7B6.744.333.1
Llama 3.1 8B7.562.437.0
Gemma 2 9B7.668.743.8
Ministral 8B8.370.941.3

Code and Math

ModelMBPP (pass@1)HumanEval (pass@1)Math (maj@1)
Gemma 2 2B54.542.722.8
Llama 3.2 3B64.661.038.4
Ministral 3B67.777.451.7
Mistral 7B50.238.413.2
Llama 3.1 8B69.767.149.3
Gemma 2 9B68.567.747.4
Ministral 8B70.076.854.5
Model Specifications
Context Length131072
Quality Index0.45
LicenseCustom
Last UpdatedOctober 2024
Input TypeText
Output TypeText
PublisherMistral AI
Languages5 Languages