A series of recent research papers published on ArXiv has highlighted existing challenges and gaps in artificial intelligence evaluation methods, particularly for Large Language Models (LLMs) and Multimodal Large Language Models (MLLMs), underscoring the need for more robust benchmarks and coherent governance.
What happened
AISafetyBenchExplorer AISafetyBenchExplorer: A Metric-Aware Catalogue of AI Safety Benchmarks Reveals Fragmented Measurement and Weak Benchmark Governance cataloged 195 AI safety benchmarks between 2018 and 2026, revealing a fragmented ecosystem and weak governance. Safety is measured inconsistently, making it difficult to compare and evaluate progress.
Concurrently, new specific benchmarks are emerging to address precise limitations. ReactBench ReactBench: A Benchmark for Topological Reasoning in MLLMs on Chemical Reaction Diagrams demonstrated that MLLMs struggle with complex topological reasoning, such as that found in chemical reaction diagrams, going beyond the recognition of individual visual elements. Similarly, HWE-Bench HWE-Bench: Benchmarking LLM Agents on Real-World Hardware Bug Repair Tasks is the first large-scale benchmark for evaluating LLM agents in real-world hardware bug repair, featuring 417 task instances from six open-source projects, highlighting their capabilities but also their limitations in complex contexts.
Another study, "Beyond One Output" Beyond One Output: Visualizing and Comparing Distributions of Language Model Generations, pointed out that users often evaluate LLMs based on single responses, overlooking the vast distribution of possible outputs. This approach limits understanding of stochasticity and models' sensitivity to small prompt changes. In parallel, basic research continues to explore how AI can learn fundamental concepts, such as the spontaneous concatenation of words from raw speech, as described in "Basic syntax from speech" Basic syntax from speech: Spontaneous concatenation in unsupervised deep neural networks, showing progress in understanding linguistic learning mechanisms.
Why it matters
This fragmentation and lack of clear governance over AI safety benchmarks have direct implications for trust and responsible AI adoption. If we cannot consistently and reliably measure the safety and true capabilities of models, it becomes difficult for regulators, developers, and end-users to make informed decisions. Limitations in topological reasoning or hardware bug repair are not just technical problems; they indicate fundamental gaps in AI's ability to operate in critical contexts such as science, engineering, or healthcare.
The tendency to evaluate LLMs via single outputs leads to a distorted perception of their actual capabilities and vulnerabilities. This can lead to overconfidence or underestimation of risks, especially in applications where consistency and robustness are essential. Understanding output distributions is crucial for bias mitigation and for developing more reliable systems. An AI's ability to learn basic syntax from speech, while foundational, must be accompanied by a deep understanding of how these capabilities translate into reliable and safe performance in complex scenarios.
The HDAI perspective
The AI benchmark ecosystem is rapidly expanding, but its fragmentation and weak governance pose a significant obstacle to the development of truly ethical and human-centered artificial intelligence. It's not enough to create new tests; it is imperative to develop shared standards and oversight mechanisms that ensure safety, reliability, and transparency are measured uniformly and meaningfully. Without robust governance, the risk is that technical advancements will be accompanied by an uncontrolled increase in social and operational risks.
The human perspective requires that AI evaluation goes beyond purely technical metrics, considering the impact on human decision-making processes, public safety, and fairness. It is crucial that benchmarks reflect real-world scenarios and that their interpretation takes into account ethical and social implications. Only then can we build AI systems that are not only powerful but also responsible and aligned with human values.
What to watch
The debate over benchmark standardization and the creation of independent governance bodies for AI evaluation is set to intensify. It will be crucial to observe how legislative initiatives, such as the AI Act in Europe, will seek to address these challenges, promoting a more coordinated and transparent approach to the evaluation and certification of AI systems. The development of methodologies for visualizing and analyzing LLM output distributions will be equally important for a more granular understanding of their behavior.

