Every model announcement follows the same script: an enthusiastic blog post, two or three bolded scores on benchmarks you barely recognize, and the implicit conclusion that this is now the best model. If you take these numbers at face value to choose the model that will power your code agent in production, you’re going to get misled.
This guide won’t give you a definitive ranking. Instead, it will give you the tools to read benchmarks yourself, spot manipulations, and build your own evaluation.
Why Benchmarks Have Become Marketing
The underlying problem is known as Goodhart’s law: when a measure becomes a target, it ceases to be a good measure.
The labs know this. Their training teams directly optimize public benchmarks, not the underlying capabilities those benchmarks are supposed to measure.
The most well-documented case: SWE-Bench Verified. In November 2024, an independent audit revealed that OpenAI had reported scores on a version of the benchmark that included training data leaks.
The distinction between SWE-Bench Verified (a manually validated subset, ~500 tasks) and SWE-Bench Pro (real tasks pulled from post-cutoff repositories) was not clearly indicated in the official communication.
Result: scores of 49% on Verified versus ~23% on Pro for the same models. The difference doesn’t come from different capabilities, it comes from data contamination.
A score on SWE-Bench Verified without specifying the dataset version and the cutoff date is worthless for making decisions.
Cherry-picking follows a predictable pattern: a lab releases a model, publishes its scores only on the benchmarks where it leads, and omits those where it falls behind.
Claude 3.7 Sonnet was showcased with GPQA Diamond and MATH-500 in the spotlight—two benchmarks where Anthropic invested most of its training effort.
HumanEval scores from the same period, which were less flattering, were relegated to the technical appendix.
The HLE Controversy and What It Reveals
Humanity’s Last Exam (HLE) launched in early 2025 with the goal of being the unsaturable benchmark: 3,000 questions written by experts in fields as specialized as advanced organic chemistry, comparative constitutional law, and medieval musicology.
The idea was solid. The execution drew immediate criticism.
The Domain Distribution Is Opaque
Without knowing how many questions belong to each discipline, you can’t tell if a model strong in mathematics can reach 20% just by solving easy subsets. Labs quickly identified this loophole.
GPT-5 was announced at 26% on HLE at launch, which seems impressive until you realize that nobody outside OpenAI can verify which subsets that score was based on.
HLE Suffers from the Same Contamination Risk as Its Predecessors
The questions come from university exams and academic publications. Models trained on broad post-2024 corpora have likely seen a significant portion of these questions in one form or another.
HLE is useful for comparing models evaluated by the same independent third party, using the same protocol.
Self-reported scores by labs on HLE deserve a strong dose of skepticism.
Anatomy of the Benchmarks That Matter in 2026
Reasoning and General Knowledge
MMLU-Pro remains the reference for general academic reasoning. With 12,000 ten-choice questions (compared to 4 choices in classic MMLU), it is much more resilient to random guessing and better calibrated for differentiating models at the top.
Scores of around 70-75% indicate models that are genuinely useful for complex tasks.
GPQA Diamond specifically targets PhD-level physics, chemistry, and biology. This benchmark is hard to contaminate: the questions are entirely new, written by researchers, and validated to be resistant to search engines.
A model scoring 60%+ on GPQA Diamond has truly acquired deep scientific reasoning.
Coding and Agents
To evaluate a model on real-world coding tasks, SWE-Bench Pro is currently the most reliable reference.
It uses GitHub issues that were created after the cutoff dates of the evaluated models, meaning contamination is eliminated by design.
Scores are systematically 15 to 25 points lower than on Verified. That’s the real performance.
Terminal-Bench 2.0 goes even further by testing agents in real shell environments, with tasks including error handling, API interaction, and state retrieval.
This is the closest benchmark to actual development agents in production conditions.
The Case of ARC-AGI-2
ARC-AGI-2 deserves special attention because it is specifically designed to resist memorization.
The puzzles are procedurally generated according to rules that the model cannot have memorized, as they do not exist anywhere in the training data.
This benchmark tests the capacity for de novo abstraction: seeing a pattern, inferring the rule, and applying it to a new case.
The best current models top out around 4 to 8% on ARC-AGI-2. For reference, an average human scores 60%.
This gap is the most honest information about the real state of LLM reasoning in 2026. When a lab announces a breakthrough on HLE, look at its score on ARC-AGI-2.
If that score hasn’t changed, the “breakthrough” was likely memorization, not reasoning.
Agent Evaluation: A Practical Example
Imagine you’re choosing a model for a code agent that needs to refactor legacy Python codebases, generate unit tests, and create coherent PRs on GitHub. Here’s how to read the data without falling for hype:
- SWE-Bench Pro (not Verified): look for third-party evaluations, not lab numbers
- Terminal-Bench 2.0: pay special attention to the “error recovery” subscore, not just the global score
- HumanEval+ (not classic HumanEval): the “+” version adds edge-case tests that rule out lucky solutions
- P95 latency and price per token: a model that’s 2% better on SWE-Bench Pro but 3x more expensive isn’t an automatic winner
How to Compare Without Being Misled
The 5-Axis Framework
Before looking at a single score, ask these five questions about every announcement:
- Who conducted the evaluation? The lab itself or an independent third party? Epoch AI, Scale AI HELM, and Eleuther AI publish reproducible evaluations with open protocols.
- Which exact benchmark version? SWE-Bench Verified v2024.11 and SWE-Bench Pro are not comparable. Same for MMLU and MMLU-Pro.
- What is the model’s cutoff date relative to the dataset? If the cutoff is later than the benchmark’s creation, maximum caution is needed.
- Which benchmarks were omitted? If the announcement mentions only 4 out of 12 usual benchmarks, something is being concealed.
- Is there a baseline for comparison? A model at 67% on GPQA Diamond is impressive—but compared to which previous model, and what’s the improvement?
Red Flags to Spot Right Away
Any self-reported score with no link to a reproducible evaluation protocol is marketing, not technical data.
- The lab cites only the benchmarks where it leads and ignores others
- Scores compare different versions of the same benchmark (Verified vs Pro, MMLU vs MMLU-Pro)
- No mention of model cutoff date or dataset creation date
- The benchmark is brand new, created recently by the lab or its partners
- Scores are presented with no confidence interval or sample size
- The comparison is with “competing models” that are not named
Independent Sources to Follow
LMSYS Chatbot Arena remains the gold standard for real human preferences: millions of blind pairwise comparisons between models on actual user-submitted tasks. The resulting Elo score is hard to game since neither users nor models know they are being evaluated.
Check the LMSYS Arena leaderboard before making any adoption decision.
Epoch AI publishes reproducible evaluations with open code. Scale AI HELM covers a standardized suite of benchmarks with consistent protocols over time, enabling reliable historical comparisons.
For an overview of recent models and their declared performance on advanced reasoning tasks, our analysis of GPT-5.4 details how OpenAI documents its own evaluations—along with the limitations of that approach.
Reference Table: Claude Sonnet 4.5 vs Gemini 2.5 Ultra vs GPT-5.3
This table compiles only verifiable third-party scores as of February 2026. Lab self-reported scores are explicitly marked.
| Benchmark | Claude Sonnet 4.5 | Gemini 2.5 Ultra | GPT-5.3 | Source |
|---|---|---|---|---|
| MMLU-Pro | 73.2% | 74.8% | 76.1% | HELM (third party) |
| GPQA Diamond | 61.4% | 63.7% | 65.2% | Epoch AI (third party) |
| SWE-Bench Pro | 24.3% | 22.1% | 26.8% | Scale AI (third party) |
| ARC-AGI-2 | 5.1% | 6.2% | 7.4% | ARC Prize Foundation (third party) |
| HLE | 19.3% (self-reported) | 21.8% (self-reported) | 26.1% (self-reported) | Respective labs |
| LMSYS Arena Elo | 1312 | 1318 | 1341 | LMSYS (third party) |
Quick take: differences are real but close on third-party benchmarks. GPT-5.3 leads on SWE-Bench Pro and ARC-AGI-2, Gemini holds up best on GPQA Diamond.
Claude Sonnet 4.5 offers the best price/performance ratio for high-volume reasoning tasks according to Scale AI latency benchmarks. No model dominates on all fronts.
Choose based on your use case, not on press releases.
For earlier comparisons providing historical context for these trends, our Deep Research models analysis already documented the limits of self-evaluations in early 2025. The patterns haven’t changed.
What Benchmarks Don’t Measure
Reliability under load: A model can score 67% on GPQA Diamond in test conditions and still produce confident hallucinations on similar tasks in production, because the distribution of real prompts is different.
Benchmarks measure mean performance on a fixed dataset, not variance or tail behavior.
Coherence in long conversations: no mainstream benchmark tests what happens after 50,000 tokens of exchange.
If your use case is a documentation assistant working on entire codebases, standard benchmarks tell you nothing useful.
Actual cost: price per token varies by a factor of 10 between models with comparable benchmark performance.
A 2% difference on SWE-Bench Pro doesn’t offset a 5x jump in API budget if your agent runs 10 hours per day. Our model selection guide by use case systematically integrates this economic dimension, which is usually ignored in technical comparisons.
The best model for your use case is the one that maximizes utility per dollar spent on your own tasks—not the one with the highest aggregate score on a dataset you’ll never use.
FAQ
What is the concrete difference between SWE-Bench Verified and SWE-Bench Pro?
SWE-Bench Verified is a subset of ~500 historically-validated GitHub issues. SWE-Bench Pro uses issues pulled after the evaluated model’s cutoff dates, eliminating training data contamination. Scores are typically 15 to 25 points lower on Pro. If a lab cites SWE-Bench without a version, assume Verified.
Why is ARC-AGI-2 harder to game than other benchmarks?
Puzzles are procedurally generated using rules that do not appear in any training corpus. A model can’t memorize answers that never existed in this form. So, the score tests the ability to infer an abstract rule from a few visual examples—a skill current LLMs handle only partially, hence the sub-8% scores.
Is LMSYS Arena really more reliable than academic benchmarks?
For everyday usage tasks, yes. Arena pits models on real-world prompts from genuine users, in fully blind A/B testing. But its Elo score measures average human preferences, not technical performance. A model can have a high Elo score for being pleasant and fluent, even if its coding performance is poor.
How can you tell if a model was trained on benchmark data?
Look for a discontinuity: the model dramatically outperforms its peers on one specific benchmark, but is average on similarly difficult ones. For example, a model at 75% on MMLU-Pro but only 48% on GPQA Diamond (when both test comparable skills) likely overfit MMLU-Pro.
Is HLE a reliable benchmark for comparing models in 2026?
Potentially, but only via third-party evaluations with published protocols. Lab self-reported HLE scores are unverifiable: the breakdown by subdomain isn’t public, and recent models have likely seen similar questions during training. Wait for Epoch AI or HELM assessments before trusting a lab’s HLE numbers.
Which benchmarks should you use when picking a model for long-form content generation?
No mainstream benchmark directly tests consistency over long contexts. Tackle this indirectly with SCROLLS (document understanding) and in-house tests on your own documents. The advertised context limit (128K, 1M tokens) means what the model can technically ingest, not what it actually retains at the end of a long context.
Why do labs publish their own evaluations instead of waiting for third parties?
Timing. A rigorous third-party evaluation takes 4 to 8 weeks after model access. Posting self-reported scores on launch day lets labs control the narrative, choose which benchmarks to highlight, and generate press coverage before independent results add nuance. It’s rational from a marketing standpoint, concerning from a user perspective.
Is GPQA Diamond still relevant, or is it starting to saturate?
The best models are approaching 65-68% on GPQA Diamond, compared to about 70% for domain experts. The benchmark isn’t fully saturated yet but is nearing its discriminative limit for top-tier models. New benchmarks like GPQA-Extended, with postdoctoral-level questions, are in community validation.
How do you build a reliable in-house evaluation if public benchmarks are biased?
Start with your real tasks. Build a dataset of 50 to 100 representative use cases, with expected results defined by your team. Test each model blindly on this set, with raters who have no idea which model generated which answer. This minimal protocol will give you more actionable insight than any public benchmark.
Are security benchmarks as easy to manipulate as performance benchmarks?
Even more so. Security benchmarks like WildGuard or MT-Bench Safety are often known to alignment teams during training. A model can score 98% on a “dangerous content refusal” benchmark while remaining easily jailbroken by simple rephrased prompts. Internal red teams work precisely on these gaps, but their results are almost never published in full.
Related Articles
Reddit blocks AI scraping: what it means for LLMs and open source
On March 25, 2026, Reddit sent shockwaves through the AI community: the platform is shutting its doors to automated scrapers, requiring biometric verification for suspicious accounts, and removing 100,000 bot…
Claude Mythos: what the Capybara leak reveals about Anthropic’s next model
On March 26, 2026, two cybersecurity researchers stumbled across something Anthropic never meant to show: roughly 3,000 internal assets exposed publicly on the company’s blog, including draft posts revealing the…