DeepSeek-R1-0528
Version: 1
Learn more: [original model announcement ]
The DeepSeek R1 model has undergone a minor version upgrade, with the current version being DeepSeek-R1-0528. In the latest update, DeepSeek R1 has significantly improved its depth of reasoning and inference capabilities by leveraging increased computational resources and introducing algorithmic optimization mechanisms during post-training. The model has demonstrated outstanding performance across various benchmark evaluations, including mathematics, programming, and general logic. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro.
Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks. For instance, in the AIME 2025 test, the model’s accuracy has increased from 70% in the previous version to 87.5% in the current version. This advancement stems from enhanced thinking depth during the reasoning process: in the AIME test set, the previous model used an average of 12K tokens per question, whereas the new version averages 23K tokens per question.
Beyond its improved reasoning capabilities, this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding.
Usage Recommendations
We recommend adhering to the following configurations when utilizing the DeepSeek-R1-0528 series models, including benchmarking, to achieve the expected performance:- Avoid adding a system prompt; all instructions should be contained within the user prompt.
- For mathematical problems, it is advisable to include a directive in your prompt such as: "Please reason step by step, and put your final answer within \boxed{}."
- When evaluating model performance, it is recommended to conduct multiple tests and average the results.
Model alignment
Microsoft and external researchers have found Deepseek R1 to be less aligned than other models -- meaning the model appears to have undergone less refinement designed to make its behavior and outputs more safe and appropriate for users -- resulting in (i) higher risks that the model will produce potentially harmful content and (ii) lower scores on safety and jailbreak benchmarks. We recommend customers use Azure AI Content Safety in conjunction with this model and conduct their own evaluations on production systems.Reasoning outputs
The model's reasoning output (contained within the tags) may contain more harmful content than the model's final response. Consider how your application will use or display the reasoning output; you may want to suppress the reasoning output in a production setting.Content filtering
When deployed via Azure AI Foundry, prompts and completions are passed through a default configuration of Azure AI Content Safety classification models to detect and prevent the output of harmful content. Learn more about Azure AI Content Safety . Configuration options for content filtering vary when you deploy a model for production in Azure AI; learn more .Extract from the original model evaluationFor all our models, the maximum generation length is set to 64K tokens. For benchmarks requiring sampling, we use a temperature of
$0.6
, a top-p value of 0.95
, and generate 16 responses per query to estimate pass@1.
Category | Benchmark (Metric) | DeepSeek R1 | DeepSeek R1 0528 |
---|---|---|---|
General | |||
MMLU-Redux (EM) | 92.9 | 93.4 | |
MMLU-Pro (EM) | 84.0 | 85.0 | |
GPQA-Diamond (Pass@1) | 71.5 | 81.0 | |
SimpleQA (Correct) | 30.1 | 27.8 | |
FRAMES (Acc.) | 82.5 | 83.0 | |
Humanity's Last Exam (Pass@1) | 8.5 | 17.7 | |
Code | |||
LiveCodeBench (2408-2505) (Pass@1) | 63.5 | 73.3 | |
Codeforces-Div1 (Rating) | 1530 | 1930 | |
SWE Verified (Resolved) | 49.2 | 57.6 | |
Aider-Polyglot (Acc.) | 53.3 | 71.6 | |
Math | |||
AIME 2024 (Pass@1) | 79.8 | 91.4 | |
AIME 2025 (Pass@1) | 70.0 | 87.5 | |
HMMT 2025 (Pass@1) | 41.7 | 79.4 | |
CNMO 2024 (Pass@1) | 78.8 | 86.9 | |
Tools | |||
BFCL_v3_MultiTurn (Acc) | - | 37.0 | |
Tau-Bench (Pass@1) | - | 53.5(Airline)/63.9(Retail) |
Model Specifications
Context Length128000
Quality Index0.87
LicenseMit
Last UpdatedMay 2025
Input TypeText
Output TypeText
PublisherDeepSeek
Languages2 Languages
Related Models