Anthropic just released Claude Opus 4.5 - here's how it stacks up against other leading models
Briefly

Anthropic just released Claude Opus 4.5 - here's how it stacks up against other leading models
"AI startup Anthropic released its latest model, Claude Opus 4.5, on Monday, describing it in a company blog post as "a step forward in what AI systems can do, and a preview of changes to how work gets done." Also: ChatGPT's new shopping research tool is fast, fun, and free - but can it out-shop me? The new model outperforms other industry-leading apps such as Google's Gemini 3 Pro and OpenAI's GPT-5.1 on coding tasks, according to Anthropic."
"Anthropic The company also wrote that the model "scored higher than any human candidate ever" on the "notoriously difficult" exam given to prospective engineering employees. The result "raises questions about how AI will change engineering as a profession," Anthropic wrote in its blog post. A version of Gemini 2.5 also recently scored top marks in the International Collegiate Programming Contest (ICPC), an internationally renowned coding competition."
Anthropic released Claude Opus 4.5, an AI model that advances creative problem-solving, agentic tasks, vision, reasoning, and math. Anthropic reports that the model outperforms competitors such as Google's Gemini 3 Pro and OpenAI's GPT-5.1 on coding tasks and scored higher than any human candidate ever on a notoriously difficult engineering exam. The model achieves state-of-the-art performance in agentic tool use and computer use and demonstrates flexible reasoning on complex problems. Test scenarios include an automated airline agent denying a basic-economy change request and handling a disgruntled customer, highlighting agentic behavior and safety considerations.
Read at ZDNET
Unable to calculate read time
[
|
]