OpenAI releases GPT-5.2 with a 38% reduction in hallucinations and superior performance on SWE-Bench Pro. Explore the new tiers: Instant, Thinking, and Pro.

OpenAI has officially released GPT-5.2 as part of a plan to regain its position as the leader in the developer ecosystem. After a "code red" to improve ChatGPT's usefulness for real-world development workflows, this release shows that the company's internal priorities have changed a lot. OpenAI's GPT-5.2 is the clearest sign yet that the company is putting reliability, reasoning, and production-ready performance ahead of experimental features. This is because Google Gemini and Anthropic Claude are getting more competitive.
The release of GPT-5.2 shows that the development pace is now faster and more aggressive. The model comes in three specialized tiers, each made for a different stage of the development lifecycle, to meet the needs of different developers:
Instant: Best for tasks that need low latency, basic queries, and quick information searches.
Thinking: The family's "reasoning engine," which is made for advanced math, logic, and architectural planning.
Pro: A high-fidelity tier for the hardest, most unclear problems where accuracy is very important.
According to both internal and external tests done by OpenAI, GPT-5.2 is the best model for professional use so far.
GDPval Performance: In a test that compared AI to human professionals in 44 different fields, GPT-5.2 Thinking did as well or better than human experts in more than 70% of tasks, such as operations planning and financial modeling.
SWE-Bench Pro: GPT-5.2 did better than both GPT-5.1 and Google's Gemini 3 Pro on this important software engineering test. It got a score of 55.6%, which shows that it is better at solving real-world coding problems in more than one programming language.
The "Codex" Edge: OpenAI also released GPT-5.2-Codex, which is specifically designed for "agentic" coding, along with the general release. It has Context Compaction, a built-in feature that lets the model handle huge refactors and migrations across large codebases without losing track of logic, which was a common problem in earlier versions.
One of the biggest changes in GPT-5.2 is that "hallucinations" have been cut down by a lot. Max Schwarzer, OpenAI's post-training lead, says that the "Thinking" model made 38% fewer hallucinations than GPT-5.1. This improvement is very important for business teams that need to be sure of the facts in production settings.
OpenAI is putting a lot of effort into defensive cybersecurity for the first time. GPT-5.2-Codex has advanced features for finding security holes and suggesting safe design patterns. OpenAI says that these features haven't yet reached a "high-risk" level in their Preparedness Framework. However, security experts are already testing the model to speed up defensive research.