OpenAIDevs Says GPT-5.3-Codex Is Now Available to All Developers in the Responses API
Original: GPT-5.3-Codex is now available for all developers in the Responses API. View original →
What was announced on X
In an official post on 2026-02-24, OpenAIDevs said: "GPT-5.3-Codex is now available for all developers in the Responses API." The post links directly to the model documentation page at developers.openai.com/api/docs/models/gpt-5.3-codex, signaling that this is a production availability update rather than only a preview note.
The timing matters. In OpenAI's earlier launch write-up for GPT-5.3-Codex, the company described API access as something it was working to enable safely. This new X post is therefore a concrete lifecycle milestone: API availability has moved from "soon" to "available now" for all developers using the Responses API.
How OpenAI framed GPT-5.3-Codex
OpenAI's product page for GPT-5.3-Codex describes the model as a stronger agentic coding system that combines frontier coding performance with broader reasoning and professional knowledge capabilities. The same write-up states that the model is 25% faster than GPT-5.2-Codex and reports benchmark gains across coding and computer-use evaluations.
- SWE-Bench Pro (Public): 56.8%
- Terminal-Bench 2.0: 77.3%
- OSWorld-Verified: 64.7%
Those numbers are vendor-reported figures from OpenAI's launch material, but they provide useful context for why the API rollout is meaningful for engineering teams choosing a default coding model.
Why this rollout is high-signal for developers
For teams already standardized on the Responses API, this announcement reduces integration friction: the model called out in OpenAI's own benchmark and launch narrative is now directly available through the same API surface many teams already use in production. That usually translates into simpler migration planning, faster A/B evaluation cycles, and clearer cost-performance testing against existing coding-agent stacks.
Primary sources: X post, model documentation, OpenAI launch write-up.
Related Articles
The bottleneck moved from GPUs to the API layer, and OpenAI changed the transport to keep up. By adding WebSocket mode and connection-scoped caching to the Responses API, the company says agentic workflows improved by up to 40% end-to-end and GPT-5.3-Codex-Spark reached 1,000 tokens per second with bursts up to 4,000.
Why it matters: public coding benchmarks are getting less useful at the frontier, so a fresh product-side score can move developer attention fast. Cursor says GPT-5.5 is now its top model on CursorBench at 72.8% and is discounting usage by 50% through May 2.
r/LocalLLaMA pushed this post up because the “trust me bro” report had real operating conditions: 8-bit quantization, 64k context, OpenCode, and Android debugging.
Comments (0)
No comments yet. Be the first to comment!