Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Read by thought-leaders and decision-makers around the world. Phone Number: +1-650-246-9381 Email: pub@towardsai.net
228 Park Avenue South New York, NY 10003 United States
Website: Publisher: https://towardsai.net/#publisher Diversity Policy: https://towardsai.net/about Ethics Policy: https://towardsai.net/about Masthead: https://towardsai.net/about
Name: Towards AI Legal Name: Towards AI, Inc. Description: Towards AI is the world's leading artificial intelligence (AI) and technology publication. Founders: Roberto Iriondo, , Job Title: Co-founder and Advisor Works for: Towards AI, Inc. Follow Roberto: X, LinkedIn, GitHub, Google Scholar, Towards AI Profile, Medium, ML@CMU, FreeCodeCamp, Crunchbase, Bloomberg, Roberto Iriondo, Generative AI Lab, Generative AI Lab VeloxTrend Ultrarix Capital Partners Denis Piffaretti, Job Title: Co-founder Works for: Towards AI, Inc. Louie Peters, Job Title: Co-founder Works for: Towards AI, Inc. Louis-François Bouchard, Job Title: Co-founder Works for: Towards AI, Inc. Cover:
Towards AI Cover
Logo:
Towards AI Logo
Areas Served: Worldwide Alternate Name: Towards AI, Inc. Alternate Name: Towards AI Co. Alternate Name: towards ai Alternate Name: towardsai Alternate Name: towards.ai Alternate Name: tai Alternate Name: toward ai Alternate Name: toward.ai Alternate Name: Towards AI, Inc. Alternate Name: towardsai.net Alternate Name: pub.towardsai.net
5 stars – based on 497 reviews

Frequently Used, Contextual References

TODO: Remember to copy unique IDs whenever it needs used. i.e., URL: 304b2e42315e

Resources

Our 15 AI experts built the most comprehensive, practical, 90+ lesson courses to master AI Engineering - we have pathways for any experience at Towards AI Academy. Cohorts still open - use COHORT10 for 10% off.

Publication

The AI Report Card: Decoding the Benchmark Jungle
Latest   Machine Learning

The AI Report Card: Decoding the Benchmark Jungle

Author(s): Shobhit Chauhan

Originally published on Towards AI.

The AI Report Card: Decoding the Benchmark Jungle

As I was peering through the model cards of DeepSeek-R1 , Gemini 2.5 Pro and such, I stumbled upon a whole galaxy of evaluation results against benchmarks that sounded like an alphabet soup gone wild! We’re talking about the likes of MMLU, GPQA, CLUEWSC — honestly, it feels like we’re trying to assess if our AI is ready for a Mensa meeting!

Then you have the ones that sound like they’re straight out of a gladiator film or a safari : ArenaHard, AlpacaEval, Big-Bench Hard. And let’s not forget the effortlessly cool-sounding FLEURS (yes, no one was trying to fit in the words here !).
Feeling truly inspired, I decided to create my own — CHIMERA (Completely Hypothetical Index for Measuring Existential Reasoning Aptitude)!

To add a little more spice to this intellectual gumbo, you see breakdowns — Gemini is using MMLU, Llama is using MMLU 5-shot. Not to be outdone, the challenger DeepSeek R1 throws down the gauntlet with MMLU (Pass@1, MMLU-Redux, and MMLU-Pro)!

And the numbers! Oh, the glorious numbers! Considering we get a new path-breaking, paradigm-shifting model every time sun sneezes a solar-flare, we may have enough data to train an AI just on these benchmark scores to predict the scores of future models.

So, here’s my attempt (first in the series) to untangle this web, understand what these benchmarks are actually trying to tell us, and hopefully simplify it for others.

What are AI Benchmarks, Really?

At their core, AI benchmarks are standardized tests designed to measure the performance of AI models on specific tasks. Think of them like the SATs, the Olympics, or even a really intense driving test, but for algorithms. We need some way to see how smart, capable, or efficient these digital brains are, especially compared to each other or to previous versions.

The purpose is multifaceted:

  1. Tracking Progress: How far have we come? Are models actually getting better, or just better at specific, narrow tasks? Benchmarks give us a yardstick.
  2. Comparison: When Company A releases Model X and Company B releases Model Y, benchmarks (ideally) provide a relatively objective way to compare their capabilities. Who’s the current heavyweight champion of coding? Who’s the Shakespeare of text generation?
  3. Identifying Strengths and Weaknesses: No model is perfect (yet!). Benchmarks help us pinpoint where a model excels (e.g., creative writing) and where it stumbles (e.g., complex mathematical reasoning). It’s like getting a diagnostic report card. This can help guide selection of a model for any specific use-case.
  4. Guiding Research: Poor performance on certain benchmarks can signal areas needing more research and innovation. They highlight the mountains we still need to climb.
  5. Reproducibility: In science, being able to verify results is crucial. Standardized benchmarks allow different teams to test models under similar conditions, lending credibility to performance claims.

The Landscape of AI Tasks and Models: Not a Monolith!

Before diving into specific benchmarks, it’s crucial to remember that “AI” isn’t one single thing. We have different domains:

  • Natural Language Processing (NLP): Understanding and generating human language (like ChatGPT, Claude, Llama). This is where much of the current benchmark frenzy is focused.Tasks include text classification, machine translation, question answering, and text generation.
  • Computer Vision (CV): Interpreting and understanding information from images or videos .Tasks include image classification, object detection, image segmentation, and video analysis.
  • Speech Recognition: Converting spoken language into text (like Siri, Alexa).It’s used in voice assistants, dictation software, and more.
  • Reinforcement Learning (RL): Training agents to make sequences of decisions by trial and error (like game-playing AI or robotics). It’s used in robotics, game playing, and autonomous systems.
  • …and many more specialized areas!

Similarly, models come in different flavors. Are they discriminative (making predictions based on input, like classifying an email as spam) or generative (creating new content, like writing a poem)? Are they supervised (trained on labeled data) or unsupervised (finding patterns in unlabeled data)?

Each combination of task and model type often requires its own specific set of benchmarks. You wouldn’t test an elephant on its capability to climb trees, right? Likewise, you don’t evaluate an image generator using a math quiz.

Let’s start with the current hotspot — benchmarks for Natural Language Processing (NLP) often used for Large Language Models (LLMs). These often test complex reasoning and knowledge.

1. MMLU (Massive Multitask Language Understanding)

What it is: This is a big one, a veritable academic decathlon for LLMs. MMLU tests knowledge across 57 diverse subjects, ranging from history and humanities to law and medicine to STEM fields like computer science and mathematics. Questions are multiple-choice, drawn from real-world exams and textbooks (e.g., AP exams, college-level tests).

Why it matters: Its sheer breadth aims to measure general knowledge and problem-solving ability across many domains, making it a good proxy for a model’s overall “worldliness.”

The Variations:

Zero-shot: This tests the model’s ability to answer questions on topics it hasn’t been explicitly trained on, relying solely on its pre-existing knowledge and understanding of language. It’s like asking a very well-read person about a topic they’ve encountered but never studied in detail.

Few-shot (e.g., 5-shot): The model is given a few examples (in this case, 5) of questions and correct answers in the prompt before being asked the actual test question. This tests its ability to quickly adapt and learn from context (in-context learning).

Pass@1: This usually means the model gets one chance to generate the answer. Did it get it right on the first try? This is common in coding benchmarks but sometimes applied here too.

MMLU-Redux / MMLU-Pro: These are newer variants trying to address potential issues with the original MMLU, like “data contamination” (where test questions might have accidentally leaked into the training data) or aiming for even harder, more discriminative questions. They are attempts to keep the benchmark relevant and challenging. MMLU-Redux focuses on correcting errors and inconsistencies in the original MMLU dataset’s ground truth labels. MMLU-Pro, on the other hand, aims to create a more challenging and robust benchmark by increasing the difficulty and incorporating more reasoning-focused questions, as well as increasing the number of answer options.

MMLU is like asking an AI to get a liberal arts degree and multiple STEM PhDs simultaneously, assessed via multiple-choice. No pressure!

2. GPQA (Graduate-Level Google-Proof Question Answering)

What it is: If MMLU is broad, GPQA is deep. It features extremely challenging questions written by domain experts (biology, physics, chemistry) that are designed to be hard even for experts with access to Google. They often require multi-step reasoning and synthesis of complex information.

Why it matters: It pushes the boundaries of deep reasoning and knowledge application. Can the AI connect obscure dots and solve problems that stump humans even with search engines?

The Variations:

By Dataset Size/Difficulty:

The original GPQA dataset comes in three main sizes, each representing a different level of difficulty based on the agreement between expert and non-expert validators:
GPQA Extended: This is the largest set, containing all 546 questions collected.
GPQA Main: This subset consists of 448 questions where at least half of the expert validators agreed on the correct answer, and at most two-thirds of non-expert validators answered correctly. This is often considered the primary evaluation set.
GPQA Diamond: This is the most challenging subset, containing 198 questions where all expert validators answered correctly, and no more than one out of three non-expert validators answered correctly. This set is specifically designed to be very difficult even for advanced LLMs.

By Evaluation Protocol/Prompting Strategy: Variations in how GPQA is used to evaluate models:
Zero-shot: Model answers questions without prior examples, testing inherent knowledge and reasoning.
Few-shot: Model answers questions after being given a few examples, assessing learning from limited context. The number of examples (e.g., 5-shot) can vary.
Chain-of-Thought (CoT): Model is asked to show reasoning steps before the final answer, improving performance on complex tasks. Can be combined with zero-shot or few-shot (Zero-shot CoT, Few-shot CoT)
Retrieval-augmented: Model can use external knowledge retrieval (e.g., web search, knowledge base) to help answer questions, testing integration of external information.

GPQA is the AI equivalent of a Ph.D. qualifying exam where the professors are actively trying to trip you up, and Stack Overflow can’t save you.

3. Big-Bench

What it is: Big-Bench is a massive collaborative benchmark with over 200 tasks. These tasks often involve multi-step reasoning, logical deduction, and understanding nuanced human contexts. Think of it as a massive and diverse exam for AI, testing its abilities across a huge range of subjects and tasks. Big-Bench Hard (BBH) is a curated subset of 23 particularly challenging tasks from Big-Bench that were identified as being beyond the capabilities of even large LLMs at the time of its creation.

Why it matters: BBH specifically targets the known weaknesses of LLMs. Excelling here suggests genuine progress in reasoning abilities, not just pattern matching on familiar data. It’s designed to be a frontier challenge.

The Variations:

  • BIG-Bench (The Full Benchmark): This is the original and most comprehensive version, encompassing over 200 diverse tasks contributed by numerous researchers. These tasks span a wide range of topics, from linguistics and mathematics to common-sense reasoning and social biases. The sheer scale and diversity are key characteristics of this “variation.”
  • BIG-Bench Hard (BBH): BBH is a specific and important subset of the full BIG-Bench. It comprises 23 (originally) particularly challenging tasks where early large language models struggled to outperform the average human rater. BBH focuses on tasks requiring multi-step reasoning and novel skills. Think of it as the “elite squad” of the BIG-Bench tasks.
  • BIG-Bench Lite (BBL): This is a smaller, more computationally efficient subset of 24 JSON-based tasks from the full BIG-Bench. BBL was designed to provide a quicker and cheaper way to evaluate models on a diverse set of capabilities, offering a canonical measure of model performance without the computational cost of running all 200+ tasks. It’s like the “sampler platter” of BIG-Bench.
  • By Task Categories/Keywords: The tasks within BIG-Bench are often categorized by their subject matter or the type of reasoning they assess (e.g., linguistics, reasoning, mathematics, social bias). Researchers might choose to evaluate models on specific categories of tasks within BIG-Bench to focus on particular capabilities. This allows for a more granular analysis of a model’s strengths and weaknesses. It’s like saying, “Let’s see how well this model does on all the ‘logical reasoning’ tasks in BIG-Bench.”

By Evaluation Protocol/Prompting Strategy: Similar to other benchmarks, the way BIG-Bench tasks are evaluated can vary:

  • Zero-shot: Evaluating the model’s performance without any in-context examples.
  • Few-shot: Providing a small number of input-output examples to guide the model. The number of “shots” can be a variation.
  • Chain-of-Thought (CoT): Prompting the model to explicitly show its reasoning steps. This has been shown to significantly impact performance on some BIG-Bench tasks, particularly those in BBH.

BIG-Bench Extra Hard (BBEH): This is a more recent evolution, building upon the original BBH by increasing the difficulty of some tasks and introducing new, even more challenging ones. BBEH represents the ongoing effort to push the boundaries of LLM evaluation. It’s like the “next level” after conquering BBH.

BBH is like the American Ninja Warrior course for LLMs — only the most agile and powerful minds (or architectures) make it through consistently.

4. GSM-8K (Grade School Math 8K)

What it is: This benchmark focuses on mathematical reasoning, specifically multi-step word problems typically found in grade school curricula (around 8,000 of them). The problems require understanding the text, identifying the steps needed, and performing basic arithmetic operations.

Why it matters: It tests whether the model can translate natural language descriptions into mathematical procedures and execute them correctly. It’s less about complex math and more about reasoning through a problem.

The variations:

By Number of Shots (Few-shot Evaluation):

  • Zero-shot: The model is given the problem directly without any examples.
  • Few-shot (e.g., 5-shot, 8-shot): The model is given a few examples of GSM-8K problems and their solutions before being asked to solve the target problem.

With Chain-of-Thought (CoT) Prompting: Another significant variation in evaluation is the use of Chain-of-Thought prompting:

  • Standard Prompting: The model is simply asked to provide the answer.
  • CoT Prompting: The prompt is designed to encourage the model to show its reasoning steps before giving the final answer (e.g., “Let’s think step by step.”).

Augmented with External Knowledge or Tools: Some approaches involve augmenting the language model’s capabilities with external tools or knowledge bases when tackling GSM-8K problems. This isn’t a variation of the dataset itself, but a different evaluation setup:

  • Calculator Integration: Allowing the model to use a calculator for arithmetic operations. The original GSM-8K paper even explored training models to use calculator annotations.
  • Retrieval-Augmented Generation: Enabling the model to retrieve relevant information from external sources that might help in solving the math problems.

“Platinum” or Cleaned Versions: Recognizing potential noise or ambiguities in the original GSM-8K, researchers have created revised or cleaned versions of the dataset:

  • GSM8K-Platinum: This revised version aimed to reduce label noise by manually inspecting and correcting or removing problematic questions. Using this cleaned version leads to a more accurate assessment of model performance.

Meta-Reasoning Focused Variations: Some work has shifted the focus from just getting the answer right to evaluating the model’s ability to reason about the correctness of a given solution:

  • MR-GSM8K: This benchmark challenges models to predict whether a given solution to a GSM-8K problem is correct, locate the first error if it’s wrong, and explain the reason for the error. This evaluates a deeper level of understanding than just solving the problem itself.

Multilingual Variations: While the original GSM-8K is in English, there have been efforts to create multilingual versions to evaluate cross-lingual mathematical reasoning:

  • MGSM (Multilingual Grade School Math): This dataset contains grade school math word problems translated into multiple languages. Evaluating on different language subsets of MGSM can be seen as a variation related to GSM-8K.
  • GSM8K-TR: This specifically refers to the Turkish translation of the GSM-8K dataset.

Synthetic or Generated Variations: Researchers have also explored generating synthetic math word problems inspired by the structure and difficulty of GSM-8K to create larger or more controlled datasets for training or evaluation:

  • GSM-Symbolic: This benchmark generates variants of GSM-8K questions using symbolic templates, aiming to reduce data contamination and allow for more controllable difficulty levels.
  • TinyGSM: This refers to a smaller, GPT-3.5 generated synthetic dataset used in conjunction with a verifier to achieve competitive results on the original GSM-8K with smaller language models.

GSM-8K asks: Can your multi-billion parameter model figure out how many apples Johnny has left? Sometimes, the answer is surprisingly “not reliably!” It’s humbling.

5. MATH (Measuring Mathematical Reasoning)

What it is: Stepping up significantly from GSM-8K, the MATH dataset contains challenging competition mathematics problems (from contests like AMC 10/12, AIME). These require more advanced knowledge (algebra, geometry, number theory, etc.) and often intricate multi-step reasoning chains. Problems require not just calculation but deriving the solution path itself.

Why it matters: This is a serious test of deep mathematical reasoning and problem-solving skills, a domain where LLMs have historically struggled compared to their linguistic prowess.

The Variations:
By Difficulty Level:
The original MATH dataset is categorized into five difficulty levels (1–5), allowing researchers to evaluate models on subsets of varying complexity. This inherent stratification can be seen as a key variation.

By Subject Area: MATH covers a range of high school mathematics subjects like Algebra, Geometry, Number Theory, Probability, and Precalculus. Evaluating models on specific subject subsets is a way to analyze their strengths and weaknesses in different mathematical domains.

Chain-of-Thought (CoT) Prompting: Applying Chain-of-Thought prompting to MATH problems, where models are encouraged to show their step-by-step reasoning, has become a significant variation in how the benchmark is used. This often leads to substantial performance improvements and allows for a deeper analysis of the model’s mathematical reasoning process.

“Functionalized” Versions: Researchers have explored creating “functional” variants of MATH problems. Instead of static question-answer pairs, these versions involve code snippets that can generate multiple variations of the same problem with different numerical values or parameters. This aims to assess the robustness of a model’s reasoning across different instances of the same problem type.

Augmented with Step-by-Step Solutions: The original MATH dataset includes full step-by-step solutions. Variations in how these solutions are used during training or evaluation (e.g., training models to generate these steps) represent another form of variation.

Related and Inspired Benchmarks: The challenging nature of MATH has inspired the creation of other, related benchmarks that can be considered variations in a broader sense:

  • MATH-Vision (MATH-V): This benchmark focuses on mathematical reasoning within visual contexts, incorporating diagrams and images into math competition problems.
  • FrontierMath: This is a more recent benchmark designed to evaluate advanced mathematical reasoning at a level beyond typical high school competitions, targeting more complex topics.
  • UGMathBench: This benchmark focuses on undergraduate-level mathematics problems across a wider range of subjects relevant to university curricula.
  • DynaMath: This benchmark emphasizes the evaluation of mathematical reasoning robustness in Vision Language Models by using dynamically generated question variants.

If GSM-8K is elementary school homework, MATH is prepping for the International Math Olympiad. It’s where AI models prove they didn’t just learn arithmetic by rote but can actually think mathematically (or give a very convincing impression).

Benchmarks are essential tools, but they aren’t perfect. A single number rarely tells the whole story. Understanding what each benchmark measures, its limitations, and why different versions exist is crucial for interpreting those flashy scores on model cards.

Join thousands of data leaders on the AI newsletter. Join over 80,000 subscribers and keep up to date with the latest developments in AI. From research to projects and ideas. If you are building an AI startup, an AI-related product, or a service, we invite you to consider becoming a sponsor.

Published via Towards AI


Take our 90+ lesson From Beginner to Advanced LLM Developer Certification: From choosing a project to deploying a working product this is the most comprehensive and practical LLM course out there!

Towards AI has published Building LLMs for Production—our 470+ page guide to mastering LLMs with practical projects and expert insights!


Discover Your Dream AI Career at Towards AI Jobs

Towards AI has built a jobs board tailored specifically to Machine Learning and Data Science Jobs and Skills. Our software searches for live AI jobs each hour, labels and categorises them and makes them easily searchable. Explore over 40,000 live jobs today with Towards AI Jobs!

Note: Content contains the views of the contributing authors and not Towards AI.