Soon after the GPT-5.1 and Gemini 3 launch, Anthropic has launched its Claude Opus 4.5 model. The AI startup claims that its new model is the best in the world for coding, agents, and computer-use-related tasks.
Where does it rank?
Claude Opus 4.5 achieves an 80.9% score on SWE-bench Verified, a real-world software engineering benchmark. Notably, Opus 4.5 is the first-ever model to breach the 80% mark on SWE-bench Verified. In comparison, Google’s newly released class=”backlink” target=”_blank” href=”https://www.livemint.com/technology/tech-news/googles-gemini-3-pro-ai-is-here-how-does-it-fare-against-chatgpt-grok-and-claude-11763549359983.html” data-vars-page-type=”story” data-vars-link-type=”Manual” data-vars-anchor-text=”Gemini 3 Pro”>Gemini 3 Pro got a score of 76.2%, while OpenAI’s class=”backlink” target=”_blank” href=”https://www.livemint.com/technology/openais-new-gpt-5-1-codex-max-all-about-the-agentic-coding-model-that-can-work-for-long-hours-11763869499126.html” data-vars-page-type=”story” data-vars-link-type=”Manual” data-vars-anchor-text=”GPT-5.1 Codex Max”>GPT-5.1 Codex Max got a score of 77.9%.

![[CITYPNG.COM]White Google Play PlayStore Logo – 1500×1500](https://startupnews.fyi/wp-content/uploads/2025/08/CITYPNG.COMWhite-Google-Play-PlayStore-Logo-1500x1500-1-630x630.png)