DeepSeek-R1-0528
DeepSeek-R1-0528
Version: 1
DeepSeekLast updated December 2025
The DeepSeek R1 0528 model has improved reasoning capabilities, this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding.
Reasoning
Coding
Agents

Direct from Azure models

Direct from Azure models are a select portfolio curated for their market-differentiated capabilities:
  • Secure and managed by Microsoft: Purchase and manage models directly through Azure with a single license, consistent support, and no third-party dependencies, backed by Azure's enterprise-grade infrastructure.
  • Streamlined operations: Benefit from unified billing, governance, and seamless PTU portability across models hosted on Azure - all part of Microsoft Foundry.
  • Future-ready flexibility: Access the latest models as they become available, and easily test, deploy, or switch between them within Microsoft Foundry; reducing integration effort.
  • Cost control and optimization: Scale on demand with pay-as-you-go flexibility or reserve PTUs for predictable performance and savings.
Learn more about Direct from Azure models .

Key capabilities

About this model

Compared to the previous version, the upgraded model shows

Responsible AI considerations

Safety techniques

Microsoft and external researchers have found Deepseek R1 to be less aligned than other models -- meaning the model appears to have undergone less refinement designed to make its behavior and outputs more safe and appropriate for users -- resulting in (i) higher risks that the model will produce potentially harmful content and (ii) lower scores on safety and jailbreak benchmarks. We recommend customers use Azure AI Content Safety in conjunction with this model and conduct their own evaluations on production systems. The model's reasoning output (contained within the tags) may contain more harmful content than the model's final response. Consider how your application will use or display the reasoning output; you may want to suppress the reasoning output in a production setting. When deployed via Microsoft Foundry, prompts and completions are passed through a default configuration of Azure AI Content Safety classification mo

Quality and performance evaluations

Source: DeepSeek For all our models, the maximum generation length is set to 64K tokens. For benchmarks requiring sampling, we use a temperature of $0.6, a top-p value of 0.95, and generate 16 responses per query to estimate pass@1.
CategoryBenchmark (Metric)DeepSeek R1DeepSeek R1 0528
General
MMLU-Redux (EM)92.993.4
MMLU-Pro (EM)84.085.0
GPQA-Diamond (Pass@1)71.581.0
SimpleQA (Correct)30.127.8
FRAMES (Acc.)82.583.0
Model Specifications
Context Length128000
Quality Index0.87
LicenseMit
Last UpdatedDecember 2025
Input TypeText
Output TypeText
ProviderDeepSeek
Languages2 Languages