
Today, OpenAI announced GPT-5.3-Codex, a new version of its Frontier coding model that will be available via command line, IDE extension, web interface, and new macOS desktop app. (No API access yet, but it’s coming.)
According to company testing, the GPT-5.3-codecs outperforms the GPT-5.2-codecs and GPT-5.2 in SWE-Bench Pro, Terminal-Bench 2.0, and other benchmarks.
There are already some headlines out there stating that “Codex built itself”, but let’s check the reality, because this is an exaggeration. The domains used here by OpenAI are similar to those you see now in some other enterprise software development firms: managing deployment, debugging, and handling test results and evaluation. There are no claims here that the GPT-5.3-codecs themselves made.
Instead, OpenAI says that the GPT-5.3-codec “was helpful in creating itself.” You can read more about what this means in the company’s blog post.
But that’s part of the pitch with this model update – OpenAI is trying to position Codex as a tool that does more than generate lines of code. The goal is to make it useful for “all tasks across the software lifecycle – debugging, deployment, monitoring, writing PRDs, copy editing, user research, testing, metrics, and more.” Emphasis is also placed on operating the model mid-task and frequently updating the state.
<a href