Why Do We Need Benchmarks?
When a new AI model launches, the announcement almost always includes a table of numbers. You might see something like "94.2% on MMLU" or "92.1% on HumanEval" splashed across a blog post or press release. These numbers are benchmarks, and they are the closest thing we have to a standardized way of measuring how capable an AI system is.
But here is the problem: unlike measuring the speed of a car or the resolution of a camera, measuring the "intelligence" of an AI system is extraordinarily difficult. There is no single number that captures everything a model can do. Benchmarks are imperfect tools, and understanding their strengths and limitations is essential if you want to make sense of AI news.
Think of benchmarks like standardized tests for students. A high SAT score tells you something about a student's abilities, but it does not tell you whether they are creative, whether they work well in teams, or whether they will succeed in a specific career. AI benchmarks work similarly. They measure specific capabilities under specific conditions, and they leave a lot out.
The Major Benchmarks You Will See in Headlines
MMLU: The General Knowledge Exam
MMLU stands for Massive Multitask Language Understanding. It is essentially a giant multiple-choice exam covering 57 different subjects, from elementary mathematics to professional law, from clinical knowledge to college-level physics. There are roughly 16,000 questions in total.
When someone says a model scores 90% on MMLU, they mean it answered 90% of these questions correctly. The test was designed to measure how much factual knowledge a model has absorbed and how well it can reason about that knowledge.
Why it matters: MMLU became one of the most widely cited benchmarks because it covers such a broad range of topics. It gives a rough sense of a model's general knowledge.
Why it is limited: Multiple-choice questions are not how most people actually use AI. You rarely ask ChatGPT to pick between options A through D. Real-world tasks are open-ended, messy, and context-dependent. A model can score very well on MMLU and still struggle with a nuanced conversation or a creative writing task.
HumanEval: The Coding Test
HumanEval is a benchmark specifically for code generation. It presents the model with 164 programming problems, each described in plain language, and asks the model to write Python code that solves the problem. The code is then actually run against test cases to see if it produces correct results.
For example, a problem might ask: "Write a function that takes a list of numbers and returns the second-largest number." The model writes code, and that code is tested automatically.
Why it matters: As AI coding assistants become more common, knowing how well a model can write working code is genuinely useful information.
Why it is limited: 164 problems is a small sample. The problems are also relatively self-contained, like homework assignments. Real software engineering involves understanding large codebases, dealing with ambiguous requirements, and debugging complex systems. HumanEval does not test any of that.
GPQA: The Expert Science Exam
GPQA, which stands for Graduate-Level Google-Proof Questions in Adversarial settings, is a benchmark designed to be genuinely hard. The questions were written by domain experts in physics, chemistry, and biology, and they were specifically crafted so that you cannot easily find the answer by searching the internet.
The name "Google-Proof" is key. These questions require deep reasoning and specialized knowledge, not just pattern matching or recall.
Why it matters: As models get better, easier benchmarks become less useful because every model scores above 90%. GPQA helps distinguish between models at the frontier of capability.
Why it is limited: It covers only science subjects, so it tells you nothing about a model's literary analysis, ethical reasoning, or practical problem-solving abilities.
MATH: The Problem-Solving Test
The MATH benchmark contains 12,500 competition-level mathematics problems drawn from high school math competitions like AMC, AIME, and Olympiad-style contests. These are not simple arithmetic questions. They require multi-step reasoning, creative problem-solving, and the ability to apply mathematical concepts in novel ways.
Why it matters: Mathematical reasoning is often considered a litmus test for genuine understanding versus memorization. A model that can solve novel math problems is demonstrating something beyond simple recall.
Why it is limited: Competition math is a very specific skill. Being good at it does not necessarily mean a model can help you with data analysis, financial modeling, or the kind of math that comes up in everyday life.
ARC: The Reasoning Challenge
ARC, the AI2 Reasoning Challenge, focuses on grade-school level science questions. But do not let "grade school" fool you. The challenge version of ARC contains questions that require genuine reasoning, not just retrieval of facts. A question might require combining two pieces of knowledge that are never stated together, similar to how a student needs to think through a problem rather than just recall an answer.
Why it matters: It tests reasoning at a fundamental level, which is arguably more important than raw knowledge.
Why it is limited: Like all multiple-choice tests, it constrains the model's response in ways that do not reflect real use.
What Leaderboards Show (and What They Hide)
If you visit sites like the Hugging Face Open LLM Leaderboard, you will find models ranked by their scores on various benchmarks. These leaderboards can be useful, but they also create misleading impressions.
What leaderboards show
Leaderboards give you a snapshot of relative performance on specific tasks. They make it easy to see that Model A scored higher than Model B on a particular benchmark. They also track progress over time, showing how quickly the field is advancing.
What leaderboards hide
Real-world performance varies. A model that tops the leaderboard on MMLU might be mediocre at following complex instructions in a conversation. Benchmarks test narrow capabilities, but users care about the overall experience.
Speed and cost are invisible. A model might score two points higher on a benchmark but take three times as long to respond and cost five times as much to run. Leaderboards typically do not show these tradeoffs.
The feel of a model is hard to quantify. Some models produce more natural-sounding text, handle ambiguity better, or are more helpful in their responses. These qualities are hard to capture in any automated test.
Context window is not tested. Most benchmarks use short prompts. They do not test how well a model handles very long documents, maintains coherence over extended conversations, or manages complex multi-turn interactions.
Benchmark Gaming and Contamination
Here is one of the most important things to understand about AI benchmarks: there are strong incentives to game them.
What is benchmark contamination?
Benchmark contamination happens when a model has seen the actual benchmark questions, or very similar questions, during its training. Imagine taking a final exam where you have already seen all the questions beforehand. You might score perfectly, but that score does not reflect your actual understanding of the material.
AI models are trained on massive datasets scraped from the internet. Benchmark questions and their answers are often available online. If a model has been trained on data that includes the benchmark itself, its high score might be meaningless.
This is not always intentional. A company might scrape a huge chunk of the internet for training data without realizing that benchmark questions were included. But intentional or not, the result is the same: inflated scores that do not reflect genuine capability.
How companies game benchmarks
Beyond contamination, there are subtler ways to make benchmark numbers look better:
Selective reporting. A company might run their model on twenty benchmarks and only report the five where it performed best. If you only see cherry-picked results, you get a distorted picture.
Benchmark-specific tuning. Models can be fine-tuned specifically to perform well on certain benchmarks. This is like teaching to the test. The model gets better at the benchmark without necessarily getting better at the underlying skill.
Evaluation tricks. The way you format the prompt, the number of examples you provide, and various technical details can all affect scores. Two companies might report scores on the "same" benchmark but with different evaluation setups, making direct comparison misleading.
Chatbot Arena and Human Preference
Because traditional benchmarks have so many limitations, researchers have developed alternative approaches. The most influential is Chatbot Arena, developed by the LMSYS group at UC Berkeley.
How Chatbot Arena works
The concept is simple but powerful. A user submits a prompt, and two anonymous models generate responses. The user then votes for which response they prefer, without knowing which model produced which answer. After thousands of these head-to-head comparisons, models receive an Elo rating, similar to how chess players are ranked.
Why it is valuable
Chatbot Arena measures what benchmarks cannot: which model do actual humans prefer for actual tasks? It captures the overall quality of responses, including factors like helpfulness, accuracy, writing quality, and instruction following that are hard to test with automated benchmarks.
Because users bring their own prompts, the test covers a natural distribution of real-world tasks rather than a predefined set of questions. And because the comparison is blind, it is harder to game.
Its limitations
Chatbot Arena has its own issues. The users who participate skew toward tech-savvy English speakers, so the ratings might not reflect how well models serve other populations. Short responses and flashy formatting can sometimes win votes even when a more measured response would be more accurate. And because it requires human judgment, it is slower and more expensive to scale than automated benchmarks.
How to Read Benchmark Comparisons Critically
Given everything we have discussed, here is a practical guide for interpreting benchmark numbers when you encounter them in the news.
Ask what the benchmark actually tests
When you see a headline like "Model X achieves state-of-the-art on GPQA," ask yourself: what does that benchmark measure? Is it testing the kind of capability you actually care about? A model that excels at graduate-level science questions might not be the best choice for writing marketing copy or summarizing legal documents.
Look for multiple benchmarks
A model that performs well across many different benchmarks is more likely to be genuinely capable than one that tops a single leaderboard. Consistency across diverse tests is a better signal than any individual score.
Check the gap, not just the ranking
There is a big difference between a model that scores 95% versus 70% and one that scores 95% versus 94.5%. Headlines often emphasize rankings ("Model X beats Model Y!") without noting that the actual difference is within the margin of error. A half-point improvement on a benchmark might be statistically insignificant or practically meaningless.
Consider the source
Who is reporting the benchmark results? If a company is reporting results for its own model, take the numbers with a grain of salt. Independent evaluations by third parties tend to be more reliable. Pay special attention to evaluations where the company had no control over the testing process.
Look for real-world evaluations
The most informative assessments often come from domain experts testing models on their actual work. A lawyer evaluating AI legal analysis, a programmer testing AI code generation in their daily workflow, or a teacher assessing AI tutoring capabilities can provide insights that no benchmark can match.
Remember that benchmarks are a snapshot
A benchmark score tells you how a model performed on a specific set of questions at a specific point in time. Models can be updated, fine-tuned, and improved. The model you use today might perform differently than the version that was benchmarked six months ago.
The Future of AI Evaluation
The field of AI evaluation is itself rapidly evolving. Researchers are developing new benchmarks that are harder to game, more reflective of real-world use, and better at testing the capabilities that matter most.
Some promising directions include:
Dynamic benchmarks that continuously generate new questions, making contamination impossible. If the questions change every time, a model cannot have memorized the answers.
Task-based evaluations that test whether a model can actually complete a useful task, not just answer questions about it. Can the model write a working application, not just solve isolated coding problems? Can it conduct a genuine research analysis, not just answer multiple-choice questions about research methods?
Adversarial evaluations designed by red teams specifically trying to find weaknesses. These evaluations test the boundaries of model capability rather than measuring performance on typical inputs.
The key takeaway is this: benchmarks are useful tools, but they are only tools. They provide one lens through which to evaluate AI systems, and they should always be supplemented with your own experience and judgment. When you see benchmark numbers in a headline, you now have the context to understand what they mean and, just as importantly, what they do not.
See This in the News
Benchmarks are not just academic exercises. They shape how companies position their products and how the public perceives AI progress. For a real-world example of how benchmark results make headlines, see how legal-domain benchmarks were used to evaluate Claude's reasoning capabilities: Claude Legal Reasoning Benchmarks. Notice how domain-specific benchmarks can tell a very different story than general-purpose ones, and consider what those results reveal about the model's strengths and limitations in specialized fields.