Anthropic's Claude Opus 4.7 Edges Out AI Rivals
Anthropic has publicly released Claude Opus 4.7, its most powerful generally available large language model, surpassing OpenAI's GPT-5.4 and Google's Gemini 3.1 Pro on key benchmarks including agentic coding, tool-use, and financial analysis. It leads the market on the GDPVal-AA knowledge work evaluation with an Elo score of 1753, compared to GPT-5.4's 1674 and Gemini 3.1 Pro's 1314.
However, the overall lead is narrow, with Opus 4.7 beating GPT-5.4 by just 7-4 on directly comparable benchmarks. Anthropic continues to restrict its even more powerful model, Mythos, to a select group of enterprise partners for cybersecurity testing purposes.
