New AI Frontiers: From Robotics to Math, the Challenge of Robust Evaluation
Artificial intelligence research is experiencing unprecedented acceleration, with advancements touching diverse fields such as robotics, occupational prediction, and mathematical discovery. However, this rapid evolution raises a crucial question: how can AI systems be evaluated effectively and responsibly, especially when operating in complex contexts with direct impacts on human lives? The need for robust metrics and a deep understanding of model behavior is fundamental to ensuring ethical AI and reliable systems.
What happened
Recent studies published on ArXiv highlight the breadth of current research directions. In the field of human-machine interaction and data analysis, a new benchmark called SQLyzr SQLyzr: A Comprehensive Benchmark and Evaluation Platform for Text-to-SQL has been introduced to more thoroughly evaluate Text-to-SQL models. These models, which allow querying databases using natural language, have seen significant improvements thanks to Large Language Models (LLMs). SQLyzr aims to overcome the limitations of evaluations based on a single aggregate score, offering a more comprehensive platform that considers different query types and realistic scenarios.
Another study explores the use of LLMs for next occupation prediction On Reasoning Behind Next Occupation Recommendation. The authors developed an innovative approach based on a "reason generator" that, by analyzing a user's educational and career history, summarizes their preferences to then feed an occupation predictor. This two-step system seeks to better align LLMs with career paths, an area where traditional models still show gaps.
In the robotics sector, Vision-Language-Action models (VLA) are demonstrating remarkable capabilities in complex applications. Research titled "How VLAs (Really) Work In Open-World Environments" How VLAs (Really) Work In Open-World Environments examines how these models perform in real-world environments and long-horizon tasks, such as household chores. The study criticizes current metrics, often based only on final success or partial scores, emphasizing the need for evaluations that consider the entire process and not just the final state of objects.
Finally, AI is opening new avenues in fundamental scientific research. An example is the application of SAT solvers and LLM-generated code for mathematical discovery in the field of Ramsey graphs Doubly Saturated Ramsey Graphs: A Case Study in Computer-Assisted Mathematical Discovery. This human-machine collaboration has led to the identification of infinite families of graphs, answering a question unresolved since 1982, and formalizing correctness proofs. Research on Kolmogorov-Arnold Networks (KANs) Scaling of Gaussian Kolmogorov--Arnold Networks, a promising neural network architecture, also continues to explore the parameters influencing its behavior.
Why it matters
These advancements demonstrate the increasing pervasiveness of AI and its potential to transform key sectors of our society. The ability to query databases in natural language, for example, democratizes access to business information, but requires systems to be accurate and reliable to avoid misinterpretations that could lead to poor decisions. Similarly, AI in occupation prediction can offer valuable tools for career guidance, but raises crucial questions about algorithmic transparency and the risk of perpetuating or amplifying existing biases in the labor market. An opaque occupational recommendation system could negatively impact the AI future of work for millions of people, limiting their opportunities.
In robotics, the use of VLAs in "open-world" environments means that AI systems will increasingly interact directly with people and unstructured spaces. Their evaluation cannot be limited to achieving a final goal but must consider the safety, adaptability, and robustness of behavior in unforeseen situations. AI-assisted mathematical discovery, finally, highlights AI's potential as an intellectual enhancement tool, but requires rigorous validation of results and an understanding of how AI arrives at its conclusions. The common challenge is to ensure that innovation is accompanied by a deep understanding and an ethical evaluation of its impact.
The HDAI perspective
The fragmented approach to AI evaluation, often limited to aggregate performance metrics or ideal scenarios, is no longer sufficient. The vision of Human Driven AI (HDAI) emphasizes that technological advancement must go hand-in-hand with careful consideration of human and social implications. These recent studies reinforce the urgency of developing holistic evaluation methodologies that include not only technical accuracy but also the robustness, transparency, fairness, and ethical impact of AI systems. It is crucial that AI governance equips itself with tools capable of scrutinizing the "why" behind AI decisions, not just the "what." Topics such as the need for more realistic benchmarks (SQLyzr), understanding LLM reasoning (occupation prediction), and analyzing robot behavior in real contexts (VLAs) will be central to discussions at the HDAI Summit 2026 in Pompeii. Research must aim towards creating AI that is not only capable but also understandable and responsible, placing the individual at the center of the design and evaluation process.
What to watch
The evolution of evaluation methods will be crucial. We expect to see an increasing emphasis on metrics that measure not only performance but also the safety, fairness, and explainability of models. Collaboration among researchers, ethicists, and policymakers will be essential to define global standards that can guide responsible AI development.

