Benchmark Reporting as Infrastructure: What Community Evals Gets Right
Source: huggingface
A benchmark score means nothing without the conditions under which it was produced. MMLU at 91% could come from a zero-shot evaluation, a five-shot evaluation, a custom system prompt that hints at the task format, or a training run that included the test set. The number looks the same in all cases. This is the core problem Community Evals, launched by Hugging Face on February 4, 2026, is trying to address: not whether benchmarks are well-designed, but whether scores reported against them are comparable at all. Looking back a month after its release, the initiative is a quieter kind of infrastructure work than the surrounding discourse about benchmark gaming would suggest.
The Metadata Problem
MMML has saturated above 91%, and GSM8K sits above 94% across frontier models. The standard response to saturation is to replace those benchmarks with harder ones, which is why GPQA, MMLU-Pro, and HLE have taken on more prominence. But harder benchmarks run into the same metadata problem: two teams can run the same model on the same benchmark and report different numbers depending on how they configured the evaluation.
The differences are not subtle. Few-shot count changes scores significantly on most benchmarks. System prompt wording shifts multiple-choice performance. Whether you include chain-of-thought reasoning changes results on math tasks. Temperature settings matter for stochastic tasks. None of this is usually disclosed in a model card’s benchmarks table, which means every score comparison implicitly assumes an identical setup that almost certainly did not exist.
Community Evals addresses this by standardizing how scores are stored and what metadata they carry. Evaluation results go into a .eval_results/ folder in the model’s repository on the Hub, stored as YAML files:
- dataset:
id: Idavidrein/gpqa
task_id: gpqa_diamond
revision: abc123
value: 0.412
date: "2025-01-15"
source:
url: https://huggingface.co/spaces/SaylorTwift/smollm3-mmlu-pro
name: Eval traces
user: SaylorTwift
notes: "no-tools"
Even in its minimal form, this YAML is richer than a row in a leaderboard table. It links to a source, records a date, and can reference a specific dataset revision, which matters when benchmark datasets are updated and old scores become incomparable with new ones.
The Inspect AI Connection
The technical backbone of Community Evals is Inspect AI, the evaluation framework developed by the UK AI Safety Institute. Benchmarks that register with the system define their evaluation specification in an eval.yaml file using Inspect AI’s format. The HLE eval.yaml, for example, defines the full evaluation pipeline:
name: Humanity's Last Exam
evaluation_framework: "inspect-ai"
tasks:
- id: hle
config: default
split: test
field_spec:
input: question
input_image: image
target: answer
solvers:
- name: system_message
args:
template: |
Your response should be in the following format:
Explanation: {your explanation for your answer choice}
Answer: {your chosen answer}
Confidence: {your confidence score between 0% and 100% for your answer}
- name: generate
scorers:
- name: model_graded_fact
args:
model: openai/o3-mini
The field_spec defines what’s in the dataset. The solvers define the prompt pipeline. The scorers define how correctness is determined. When a benchmark publishes this file, it’s publishing the methodology alongside the data. The MMLU-Pro eval.yaml shows exactly which solver is used (multiple choice) and which scorer is applied (choice), which is enough information to know whether two scores were produced under comparable conditions.
This is a meaningful shift in how benchmark methodology gets communicated. Previously, reproducing a benchmark result required reading a paper or model card description carefully, finding the original evaluation codebase, and hoping nothing had changed since the results were published. With a standardized eval.yaml, the setup is machine-readable and attached to the benchmark dataset itself.
The Verification Tier
The most important technical detail in the YAML format is the verifyToken field. When an evaluation runs through HF Jobs using the Inspect AI backend, it generates a cryptographic token that links the stored score back to the actual computation. Including that token in the YAML file gives the result a “verified” badge on the model page.
Without the token, results are self-reported. That is not automatically untrustworthy, but it means a model author could publish any number. With a valid token, there is a chain of custody: the score is tied to a specific run in a specific compute environment. This creates three practical tiers of trust for any score on the Hub: verified (cryptographically attested by HF Jobs), community (submitted via PR by someone other than the model author), or self-reported (added by the model author without verification).
The community tier deserves attention. Anyone can open a PR against any model repo and add an eval result file. That PR appears on the model page with a “community” badge while it remains open. The model author can merge it, comment on it, or close it. This is the mechanism that makes the system genuinely open: you can publish a score for a model you did not train, and it will be visible to anyone who looks at that model’s page. Model authors retain the ability to close PRs they disagree with, but that action is itself visible, which changes the social dynamics around disputed scores.
Git as Audit Trail
Something the official announcement undersells is that storing scores in Git repositories makes the entire history of benchmark reporting auditable. You can see when a score was added, check whether it was changed after the fact, and notice if it was added suspiciously close to a model release. None of this is possible with a centralized leaderboard that shows only the current state of the world.
The PR mechanism adds another layer. Disputed scores flow through Git’s standard collaboration workflow: open a PR, leave a comment explaining the discrepancy, reference your own evaluation logs. This is the same process used to fix documentation errors in any open-source project, and it puts the infrastructure for handling disputes in a system most developers already understand. The Hub eval-results documentation notes that all scores are exposed via Hub APIs, which means third parties can build aggregation dashboards, anomaly detection, or historical tracking without needing special access.
What This Doesn’t Fix
Community Evals is infrastructure for reporting scores, not a solution to benchmark selection or training data contamination. A model trained on GPQA test questions can still report a high GPQA score through this system. The verifyToken proves the evaluation ran correctly; it does not prove the model was not trained on the test set. The system explicitly does not attempt to prevent this, though the thinking is that making all scores visible with timestamps and sources makes the game easier to spot.
The benchmark-reality gap also stays open. Models that score at the frontier on MMLU-Pro and HLE still struggle with reliable multi-step web browsing, consistent code generation across unfamiliar codebases, and long-horizon task completion without hallucinating intermediate steps. Better reporting of scores on academic benchmarks does not close that gap, and the announcement explicitly acknowledges it.
What it does is make the reporting layer more honest. If you want to compare two models on GPQA Diamond, you can now find scores with attached methodology, source links, and timestamps. You can filter by verified scores only, or look at what community contributors found when they ran independent evaluations. You can see if the model author published their score the day of the release or whether it was added six months later.
The Broader Pattern
Hugging Face has spent years building infrastructure that shifts AI development toward openness: model sharing, dataset hosting, collaborative model cards, lighteval for running evaluations at scale. Community Evals fits the same pattern. The organization is not trying to run the definitive evaluation of every model; it is building the pipes through which evaluation results can flow with enough metadata to be meaningful.
The benchmark allowlist is still manually curated by the HF team during the beta period, and the eval.yaml format is validated at push time. Both constraints will likely loosen as the system matures. The harder question is whether major research labs and model companies, some of which benefit from opaque reporting, will adopt the standard seriously. Transparent methodology disclosure is not uniformly in every model provider’s interest. The long-term value of Community Evals depends on how many of them use it anyway, and on whether the community-submitted results create enough social pressure to make opting out of the system look like it has something to hide.