Author: Denis Avetisyan
A new benchmark challenges artificial intelligence to actively seek and interpret data in physical systems, mirroring the iterative process of real-world scientific discovery.

Researchers introduce MaD Physics, a platform for evaluating agent-based AI’s ability to strategically gather information and make measurements under cost constraints.
Scientific discovery inherently demands navigating trade-offs between measurement fidelity and resource limitations, yet current AI benchmarks often prioritize either static knowledge or unconstrained experimentation. To address this gap, we introduce ‘MaD Physics: Evaluating information seeking under constraints in physical environments’, a new benchmark designed to evaluate agents’ ability to strategically gather informative measurements within budgetary and physical constraints. This benchmark-featuring environments governed by altered physical laws-assesses both model inference from data and planning under uncertainty, revealing shortcomings in current large language models like Gemini regarding structured exploration and data collection. Can these findings catalyze the development of more robust and scientifically grounded AI agents capable of genuine discovery?
The Inevitable Limits of Observation
Scientific progress frequently encounters a practical barrier: the inherent cost of obtaining measurements that are both numerous and accurate. A fundamental tradeoff exists because increasing the fidelity of data – reducing noise and uncertainty – typically demands more sophisticated, and therefore more expensive, instrumentation or prolonged observation times. This constraint isn’t merely financial; it also encompasses logistical hurdles and the potential for experimental disruption. Consequently, researchers often face a difficult choice between comprehensively exploring a system with limited precision, or focusing on a narrow subset with high-resolution data, ultimately impacting the scope and validity of derived models. This limitation is pervasive, influencing investigations ranging from subatomic particle physics – where detectors are extraordinarily costly – to large-scale ecological studies requiring extensive fieldwork and monitoring.
Investigating intricate systems – be it turbulent fluid dynamics, genomic regulatory networks, or even macroeconomic trends – frequently encounters a practical barrier: the escalating cost of high-fidelity data acquisition. Traditional experimental designs, predicated on exhaustive parameter sweeps and precise measurements at every step, quickly become untenable as system complexity increases. The sheer number of variables and their interactions demands an exponentially growing volume of data, pushing the limits of both computational resources and experimental budgets. This isn’t simply a matter of needing better instruments; it represents a fundamental challenge to the scientific method itself, forcing researchers to confront the reality that fully characterizing these systems with conventional techniques is often economically or logistically impossible, and necessitates innovative strategies for data collection and analysis.
The capacity to build accurate models and achieve genuine understanding of natural phenomena is fundamentally challenged by the practical constraints of measurement. Across disciplines – from charting the behavior of subatomic particles in physics to deciphering the intricate networks within biological systems – a lack of precise, cost-effective data often restricts the scope and fidelity of inquiry. This limitation isn’t simply a matter of technological inadequacy; it represents a core difficulty in translating theoretical frameworks into testable predictions. Consequently, researchers frequently rely on simplified representations or make assumptions that, while necessary for progress, inevitably introduce uncertainty and potentially obscure crucial details of the system under investigation. This pervasive challenge underscores the need for innovative strategies that can maximize the information gleaned from limited and imperfect measurements, ultimately paving the way for more robust and insightful scientific models.
Addressing the challenge of complex systems requires a shift towards active sensing, a methodology that moves beyond passively collecting data to intelligently probing a system’s behavior. Rather than exhaustively measuring every possible parameter, active sensing strategically selects inputs and observes the resulting outputs, iteratively refining its understanding with each interaction. This approach significantly reduces the experimental burden by focusing on the most informative measurements, effectively navigating the vast parameter space and identifying key relationships with far greater efficiency. The technique allows researchers to prioritize exploration of areas with high potential for discovery, offering a pathway to unlock insights that would otherwise remain hidden due to the limitations of traditional, static measurements.
MaD Physics: A Benchmark for Intelligent Exploration
MaD Physics serves as a standardized evaluation platform for intelligent agents focused on active sensing, specifically within environments characterized by limited resources. The benchmark assesses an agent’s capacity to strategically acquire information by actively controlling sensors and allocating measurement budgets. Performance is evaluated not simply on achieving a task, but on the efficiency with which an agent gathers necessary data given constraints on time, energy, or the number of measurements permissible. This necessitates agents to prioritize information gathering, balancing the cost of sensing with the resulting increase in knowledge about the environment and the task at hand.
MaD Physics incorporates a range of simulation environments built to facilitate comprehensive agent evaluation. These environments model physical systems governed by distinct principles: classical mechanics, encompassing rigid body dynamics and collisions; fluid dynamics, simulating the behavior of liquids and gases; and quantum mechanics, representing phenomena at the atomic and subatomic level. This diversity allows for testing agent performance across a broad spectrum of physical challenges and ensures that agents are not overfitted to specific, limited scenarios. The benchmark’s environments are designed to assess an agent’s ability to generalize its sensing and control strategies to previously unseen physical regimes.
MaD Physics environments employ numerical integration techniques – specifically, solvers for ordinary differential equations – to simulate the evolution of physical states over time. These methods approximate solutions to the equations of motion, allowing for the modeling of continuous physical systems with defined accuracy based on parameters like timestep size and solver tolerance. The use of numerical integration is critical because it enables the creation of deterministic and reproducible simulations, essential for rigorous performance evaluation of intelligent agents; variations in simulation results due to stochasticity are minimized, allowing for precise measurement of an agent’s ability to perceive and interact with its environment. The fidelity of the numerical integration directly impacts the realism of the simulation and, therefore, the validity of the benchmark.
MaD Physics necessitates strategic resource allocation by agents due to constrained measurement capabilities within its simulated environments. Agents are not afforded unlimited sensing; instead, they must actively decide which measurements to take, when to take them, and where to direct their sensors. This limitation forces agents to prioritize information gathering, balancing the cost of each measurement against its potential to reduce uncertainty about the system’s state. Effective performance, therefore, is not solely determined by the agent’s ability to process data, but crucially by its capacity to intelligently manage and optimize the use of limited sensing resources to maximize information gain and achieve its objectives.

Complexities Within Simulated Realities
The classical mechanics environment within the benchmark introduces complexities beyond standard Newtonian physics by incorporating anisotropic inertia and modified gravitational forms. Anisotropic inertia refers to mass distribution where inertial properties are direction-dependent, meaning an object’s resistance to acceleration varies based on the axis of applied force. Modified gravity scenarios deviate from the inverse-square law, potentially including non-constant gravitational constants or additional gravitational forces. These alterations necessitate that agents adapt their models of force and motion, moving beyond assumptions of isotropic mass distribution and standard gravitational interactions to accurately predict and react to physical phenomena within the simulation.
Fluid dynamics simulations within the benchmark utilize state-dependent forces, specifically an “alien gyroscopic forcing,” to introduce complexity beyond standard Newtonian fluid behavior. This forcing manifests as a torque applied to fluid elements proportional to their velocity and an arbitrary direction, effectively creating non-inertial frames of reference within the flow. Accurate analysis of these systems requires agents to not only measure fluid velocities and pressures, but also to strategically determine the magnitude and direction of this applied torque at each point in time and space. The state-dependent nature of the forcing means that simple extrapolation or pre-calculated models are insufficient; agents must dynamically measure the current state of the flow to infer the forces acting upon it, demanding sophisticated sensor fusion and real-time estimation techniques to accurately predict fluid behavior and achieve control objectives.
Quantum mechanics environments within the benchmark utilize non-linear entanglement, where the state of multiple particles are correlated in a manner that cannot be described by classical physics, and the Generalized Born Rule, a simplification of the full many-body Schrödinger equation used to approximate the dynamics of quantum systems. This combination necessitates that agents operating within these environments process and react to complex quantum correlations, going beyond simple pairwise interactions. Specifically, agents must infer the state of entangled particles based on incomplete or noisy measurements, and predict how these correlations will evolve under various interactions; the Generalized Born Rule introduces approximations which must be accounted for in accurate state estimation and prediction. Successful performance requires agents to effectively represent and manipulate high-dimensional quantum states and understand the implications of non-locality inherent in entanglement.
The benchmark’s design intentionally incorporates a wide range of physical simulations – including classical mechanics with modified gravitational forms, fluid dynamics with state-dependent forces, and quantum mechanical environments leveraging non-linear entanglement – to assess an agent’s capacity for generalization. Successful performance isn’t predicated on memorizing solutions to specific problems, but rather on the ability to extract underlying physical principles and apply them to novel, unseen scenarios. This necessitates the development of agents that can reason scientifically – formulating hypotheses, interpreting data from diverse simulations, and adapting their strategies based on observed outcomes – rather than relying on pre-programmed responses to defined conditions.

The Promise of LLMs in Scientific Discovery
Recent advancements showcase the remarkable potential of large language models (LLMs) to contribute directly to scientific discovery, as exemplified by the performance of Gemini models within the MaD Physics benchmark. These models aren’t simply processing data; they’re functioning as autonomous agents, actively investigating physical systems through simulated experiments. By formulating hypotheses, designing measurements, and interpreting results, the LLMs demonstrate a capacity for iterative learning and refinement of understanding – a hallmark of the scientific method. This approach moves beyond traditional data analysis, allowing the models to proactively seek information and uncover underlying principles within complex environments, suggesting a future where LLMs can assist – and perhaps even lead – in the exploration of the natural world.
The LLM-based agents don’t passively observe; instead, they actively investigate their environments using a strategy akin to a scientist formulating and testing hypotheses. This involves active sensing, where the agent strategically chooses what measurements to take, rather than simply receiving data. Crucially, these agents employ adaptive experimental design, meaning their measurement choices aren’t random, but informed by previous results. By analyzing initial data, the agent refines its approach, focusing on areas of the parameter space most likely to yield useful insights – a process that dramatically increases efficiency in discovering underlying principles. This targeted exploration allows the agent to converge on accurate models with significantly fewer trials than traditional methods, demonstrating a powerful paradigm for automated scientific discovery.
Beyond simply predicting outcomes, large language model agents operating within physics simulations enable the recovery of the governing equations themselves through the application of symbolic regression. This computational technique analyzes the agent’s collected data – the measurements taken during its active exploration of a physical system – and attempts to identify mathematical expressions that best fit the observed relationships. Essentially, it reverse-engineers the underlying physics from data, potentially revealing F = ma or other fundamental laws without prior knowledge. The ability to automatically discern these equations from agent behavior not only validates the agent’s learning process but also provides a powerful tool for scientific discovery, particularly in complex systems where analytical solutions are intractable and human intuition may fail.
Analysis of the MaD Physics benchmark reveals a nuanced relationship between agent performance, environmental complexity, and model capability. Prediction error isn’t uniform across the simulated environments; certain physical scenarios present greater challenges to accurate inference than others, influencing the agent’s ability to discern underlying principles. Critically, this study demonstrates a clear correlation between model sophistication and predictive accuracy; newer, more powerful models – specifically progressing from Gemini 2.5 Flash Lite to Flash, Pro, and ultimately Gemini 3 Flash – consistently exhibit lower prediction errors. This improvement isn’t merely incremental; the leap to Gemini 3 Flash showcases a significant advancement in learning capabilities, suggesting that continued development of large language models holds substantial promise for accelerating scientific discovery through automated experimentation and data analysis.
Recent studies utilizing the MaD Physics benchmark reveal a significant advancement in learning capabilities between large language models. Specifically, Gemini 3 Flash demonstrably outperforms Gemini 2.5 Pro within classical mechanics environments, achieving lower prediction error and a markedly faster convergence rate. This improved performance suggests that Gemini 3 Flash more efficiently identifies key relationships within the simulated physics problems, allowing it to formulate accurate predictions with fewer experimental iterations. The results, visualized in Figure 3, indicate a substantial reduction in the time required to reach a stable and accurate understanding of the underlying physical principles, positioning Gemini 3 Flash as a promising tool for automated scientific discovery and hypothesis generation.

The pursuit of scientific discovery, as highlighted by MaD Physics, isn’t merely about accumulating data, but about intelligently seeking it. This echoes Barbara Liskov’s sentiment: “Programs must be right first before they are fast.” The benchmark deliberately introduces cost-fidelity tradeoffs, forcing agents to prioritize efficient information gathering-a pragmatic approach mirroring the need for correctness before optimization. Just as a well-designed program prioritizes robust logic, these AI agents must prioritize meaningful measurements over exhaustive ones, acknowledging that architecture without history – or, in this case, informed data acquisition – is fragile and ephemeral. The true test lies not in how much data is collected, but in the quality of understanding derived from it.
What Lies Ahead?
The introduction of MaD Physics establishes a logging of capabilities-a chronicle of what current agent-based systems can discern, given finite resources. This benchmark isn’t merely a test of performance; it’s a measurement of how gracefully a system ages under constraint. Existing approaches, focused on memorization or brute-force exploration, reveal their limitations quickly. The true challenge, and the direction for future work, resides in fostering agents that prioritize information with an understanding of its inherent value-that recognize the difference between data and signal.
Deployment of these agents, this moment on the timeline, highlights a critical, unresolved problem: the translation of simulated fidelity into real-world efficacy. A system excelling in a pristine digital environment may falter when confronted with the inherent noise and ambiguity of physical systems. The cost-fidelity tradeoff, therefore, isn’t static; it’s a decaying function, demanding adaptive strategies that can recalibrate in the face of inevitable degradation.
Ultimately, MaD Physics serves as a forcing function, accelerating the development of AI that doesn’t simply process information, but seeks it-not as an end in itself, but as a means to navigate the inevitable entropy of existence. The benchmark itself will age, of course, becoming less challenging as systems improve. But the underlying question-how to extract meaningful insight from a world of limited resources-will remain perpetually relevant.
Original article: https://arxiv.org/pdf/2605.10820.pdf
Contact the author: https://www.linkedin.com/in/avetisyan/
See also:
- FRONT MISSION 3: Remake coming to PS5, Xbox Series, PS4, Xbox One, and PC on January 30, 2026
- Taylor Sheridan’s Gritty 5-Part Crime Show Reveals New Final Season Villain
- The Boys Season 5, Episode 5 Ending Explained: Why Homelander Does THAT
- Mark Zuckerberg & Wife Priscilla Chan Make Surprise Debut at Met Gala
- Invincible Season 4 Episode 4 Post-Credits Unveils a Demonic Return & More
- Anna Wintour Reacts to Rumors She Approves All Met Gala Looks
- From season 4 release schedule: When is episode 2 out on MGM+?
- Welcome to Demon School! Iruma-kun season 4 release schedule: When are new episodes on Crunchyroll?
- Why There’s No Ghosts Tonight (Nov 27) & When Season 5, Episode 7 Releases
- How to Build Water Elevators and Fountains in Enshrouded
2026-05-13 00:10