OpenAI Launches GPT-5-Codex-Mini Raising Rate Limits

OpenAI unveiled GPT-5-Codex-Mini, a compact model delivering about 4x more runs for high-throughput coding tasks. It also raises Codex rate limits and adds priority processing for paid ChatGPT plans.

openai cover

TL;DR

  • GPT-5-Codex-Mini: compact variant offering roughly 4× more usage than GPT-5-Codex with a modest capability tradeoff, aimed at simpler or repetitive coding tasks
  • Availability: selectable in the CLI and IDE extension when signed in with ChatGPT; API support coming soon
  • System will suggest switching to GPT-5-Codex-Mini when usage reaches 90% of configured limits
  • 50% higher rate limits for ChatGPT Plus, Business, and Edu plans driven by GPU efficiency and better utilization
  • Priority processing for ChatGPT Pro and Enterprise to improve responsiveness and throughput under load
  • Practical impact: mini model for scaffolding, simple refactors, and bulk transformations; higher caps and priority reduce interruptions in collaborative or classroom settings

OpenAI is rolling out three practical updates to Codex access aimed at extending usable cycles and speeding throughput. The changes include a new compact model, increased rate limits for several ChatGPT plans, and priority processing for higher-tier accounts — along with backend efficiency work to get more out of GPU resources.

GPT-5-Codex-Mini: more runs, slightly smaller model

A new offering, GPT-5-Codex-Mini, delivers roughly 4x more usage compared with GPT-5-Codex by virtue of being a more compact and cost-efficient variant. The tradeoff is a modest capability reduction relative to the full model, which positions the mini-model for simpler or repetitive coding tasks where extra throughput is more valuable than peak model capability.

  • Availability: selectable in the CLI and IDE extension when signed in with ChatGPT.
  • API support: indicated as coming soon.
  • The system will also suggest switching to GPT-5-Codex-Mini when usage hits 90% of configured limits, enabling longer uninterrupted sessions.

Higher limits and priority processing

Two account-level changes accompany the model update:

  • 50% higher rate limits for ChatGPT Plus, Business, and Edu plans, resulting from efficiency improvements and better GPU utilization.
  • Priority processing for ChatGPT Pro and Enterprise accounts to boost responsiveness and throughput under load.

These adjustments target both higher-frequency workflows and scenarios where latency or throughput matter most.

Practical implications for development workflows

The combination of a compact Codex model plus raised limits and priority processing encourages more flexible cost/performance tradeoffs:

  • GPT-5-Codex-Mini can be used for scaffolding, simple refactors, or bulk transformations where the primary goal is maximizing token throughput.
  • Higher rate caps and priority processing reduce interruptions and contention for users on paid tiers, particularly in collaborative or classroom settings where many sessions run concurrently.

API integration remains pending, so the immediate effects are most visible within the CLI and IDE extension when authenticated with ChatGPT. The announcement notes efficiency gains on the GPU side as a contributing factor to these stepped-up limits.

Continue the conversation on Slack

Did this article spark your interest? Join our community of experts and enthusiasts to dive deeper, ask questions, and share your ideas.

Join our community