codex-mini
codex-mini
Version: 2025-05-16
OpenAILast updated December 2025
codex-mini is a fine-tuned variant of the o4-mini model, designed to deliver rapid, instruction-following performance for developers working in CLI workflows. Whether you're automating shell commands, editing scripts, or refactoring repositories, Codex-Min
Multipurpose
Multilingual
Multimodal

Direct from Azure models

Direct from Azure models are a select portfolio curated for their market-differentiated capabilities:
  • Secure and managed by Microsoft: Purchase and manage models directly through Azure with a single license, consistent support, and no third-party dependencies, backed by Azure's enterprise-grade infrastructure.
  • Streamlined operations: Benefit from unified billing, governance, and seamless PTU portability across models hosted on Azure - all part of Microsoft Foundry.
  • Future-ready flexibility: Access the latest models as they become available, and easily test, deploy, or switch between them within Microsoft Foundry; reducing integration effort.
  • Cost control and optimization: Scale on demand with pay-as-you-go flexibility or reserve PTUs for predictable performance and savings.
Learn more about Direct from Azure models .

Key capabilities

About this model

codex-mini is a fine-tuned variant of the o4-mini model, designed to deliver rapid, instruction-following performance for developers working in CLI workflows.

Key model capabilities

a. Optimized for Speed: Delivers fast Q&A and code edits with minimal overhead. b. Instruction-Following: Retains Codex-1's strengths in understanding natural language prompts. c. CLI-Native: Interprets natural language and returns shell commands or code snippets. d. Long Context: Supports up to 200k-token inputs—ideal for full repo ingestion and refactoring. e. Lightweight and Scalable: Designed for cost-efficient deployment with a small capacity footprint. codex-mini supports features such as streaming, function calling, structured outputs, and image input. With these capabilities in mind, developers can leverage codex-mini for a range of fast, scalable code generation tasks in command-line environments.

Use cases

See Responsible AI for additional considerations for responsible use.

Key use cases

For developers seeking fast, reliable code generation in terminal environments, this purpose-built model offers a powerful new tool in your AI toolkit for fast, low-latency code generation in command-line environments.

Out of scope use cases

The provider has not supplied this information.

Pricing

Pricing is based on a number of factors, including deployment type and tokens used. See pricing details here.

Technical specs

The provider has not supplied this information.

Training cut-off date

The provider has not supplied this information.

Training time

The provider has not supplied this information.

Input formats

codex-mini supports features such as streaming, function calling, structured outputs, and image input.

Output formats

The provider has not supplied this information.

Supported languages

The provider has not supplied this information.

Sample JSON response

The provider has not supplied this information.

Model architecture

The provider has not supplied this information.

Long context

Supports up to 200k-token inputs—ideal for full repo ingestion and refactoring.

Optimizing model performance

The provider has not supplied this information.

Additional assets

The provider has not supplied this information.

Training disclosure

Training, testing and validation

The provider has not supplied this information.

Distribution

Distribution channels

This model is provided through the Azure OpenAI Service. codex-mini is now available via the Azure OpenAI API and Codex CLI.

More information

The following documents are applicable: Prompts and completions are passed through a default configuration of Azure AI Content Safety classification models to detect and prevent the output of harmful content. Learn more about Azure AI Content Safety . Additional classification models and configuration options are available when you deploy an Azure OpenAI model in production; learn more .

Responsible AI considerations

Safety techniques

Prompts and completions are passed through a default configuration of Azure AI Content Safety classification models to detect and prevent the output of harmful content. Learn more about Azure AI Content Safety . Additional classification models and configuration options are available when you deploy an Azure OpenAI model in production; learn more .

Safety evaluations

The provider has not supplied this information.

Known limitations

The provider has not supplied this information.

Acceptable use

Acceptable use policy

The provider has not supplied this information.

Quality and performance evaluations

Source: OpenAI The provider has not supplied this information.

Benchmarking methodology

Source: OpenAI The provider has not supplied this information.

Public data summary

Source: OpenAI The provider has not supplied this information.
Model Specifications
Context Length200000
LicenseCustom
Training DataMay 2024
Last UpdatedDecember 2025
Input TypeText,Image
Output TypeText
ProviderOpenAI
Languages1 Language