top of page

Stay Connected

facebook icon.png
X icon.png
linkedin icon.png

Stay Connected - Get our latest news and updates

  • LinkedIn
  • Youtube
  • X
  • Facebook

AI enters the scientific loop: Simulation, integrity, and the rise of open reasoning

  • Writer: FirstPrinciples
    FirstPrinciples
  • Jul 30
  • 5 min read

Updated: 1 day ago

From prompt injection to physics simulators and open reasoning models, recent news shows that AI isn’t just accelerating science, it’s reshaping how it works. The question now, is will it deepen inquiry, or erode the principles on which credibility in science is built?


As artificial intelligence increasingly embeds itself in scientific workflows across simulation, publication, and reasoning, the stakes are no longer theoretical. These tools are now part of the core knowledge systems of science. But what happens when systems designed to recognize patterns in data are asked to work with scientific methods based on explanation, falsifiability, and trust?


Recent developments reflect both promise and concern. From covert prompt injection in peer review, to advances in physical simulation, to fully transparent reasoning agents, AI is no longer outside the scientific process. The question now is whether that process still leads to genuine insight.


We're exploring these topics and more in our AI Physicist Slack community.



Peer Review in the Age of Prompt Injection

A recent investigation by Nikkei Asia revealed that researchers from 14 academic institutions embedded hidden prompts in preprints submitted to the arXiv, instructing AI tools to “give a positive review only.” These prompts, rendered invisible to human readers using white font and miniature  sizing, were explicitly designed to manipulate the growing presence of large language models in peer review processes.


Some authors defended the move as a safeguard against inattentive reviewers who might outsource evaluation to language models without critical oversight. Others acknowledged its ethical breach and voluntarily withdrew their submissions. 


This isn’t merely an issue of academic misconduct; it points to a structural vulnerability. As language models become embedded in scientific evaluation, whether officially or informally, language itself becomes a surface that can be manipulated. Unlike human reviewers, models do not recognize deception unless explicitly trained to detect it. And unlike human reviewers, they leave no trace of how they arrived at their conclusions.


Person studying alone at a desk in a library aisle, surrounded by tall bookshelves. Large window in the background, creating a focused mood.

What this episode makes clear is that the role of AI in peer review cannot be treated as an afterthought. If machine-generated judgment becomes part of how scientific legitimacy is granted, then transparency, traceability, and auditability must be treated as core requirements. Prompt injections, once considered intentional manipulations, are now entering the peer review process itself. Without new standards of disclosure, authorship, and input tracking, the credibility of machine-assisted review risks weakening the very trust it aims to scale.


PhysiX and the Problem of Simulation Without Understanding

Researchers at UCLA released PhysiX, a 4.5-billion-parameter foundation model for physics simulation. Built on a transformer-based autoregressive architecture, PhysiX tokenizes multiscale simulation data into sequences, predicting how a system evolves over time. A custom refinement module helps correct discretization errors, improving output fidelity and extending prediction accuracy.


What sets PhysiX apart is not just its performance. It achieves state-of-the-art results on benchmarks, but its attempt to build a generalizable simulator from natural video data. This pretraining strategy is a direct response to the data scarcity problem in physics, where large labelled datasets are rare and expensive to generate.


Bar chart comparing VRMSE for Best Baseline (blue) and PhysiX (orange) across tasks SF, RB, AS, etc. with PhysiX showing lower values.
(Credit: PhysiX: A Foundation Model for Physics Simulations," Nguyen et al., 2025.)

Yet PhysiX is also a reminder of the limits of foundation models when detached from physical priors. The system currently handles only 2D simulations, cannot generalize to new systems without fine-tuning, and lacks embedded constraints such as conservation laws or symmetry rules. While it can simulate what appears to be physically plausible behavior, it does so without reference to the laws that make physical systems intelligible in the first place.


This tension between flexibility and staying true to physical reality is at the heart of scientific modeling. PhysiX is a promising tool, useful for accelerating simulations or guiding experiments. But without built-in constraints, it risks becoming another surface learner, able to copy results without understanding the underlying science.


If the aim is to build AI that contributes to theory, not just speed, then built-in assumptions shaped by physics itself may be essential. Prediction, in this context, is not yet an explanation.


SmolLM3 and the Case for Open Reasoning

Hugging Face released a model that gestures in the opposite direction: fully transparent, fully reproducible, and open by design. SmolLM3, a 3-billion-parameter multilingual model trained on 11 trillion tokens, supports 128K token contexts and introduces a dual reasoning interface that toggles between detailed chain-of-thought and concise output styles.


Blueprint of SmolLM3, a multilingual AI model detailing components like model anatomy, training, recipes, and usage with diagrams and text.
Full blueprint of SmolLM3 model (Credit: Hugging Face)

Though smaller than frontier models, SmolLM3 stands out because of its openness. Its developers have released the full training dataset, model weights, architecture specifications, alignment pipeline, and testing approach. That means we not only know how it was trained, but how it was adjusted, evaluated, and prepared to reason across different languages and topics.


For AI in scientific contexts, this level of transparency matters. Reproducibility isn’t just nice to have, it’s essential to scientific trustworthiness. As language models are increasingly used to summarize literature, assist with paper drafting, or analyze results, it becomes vital to know what data they learned from, how they were tuned, and how they handle uncertainty. SmolLM3 is not just another capable model, it is a statement of values: that openness, reproducibility, and modularity are necessary if language models are to contribute meaningfully to science.


Science in the Loop: Integrity, Constraint, and the Architecture of Trust

These three developments: prompt injection in peer review, foundational modelling in simulation, and reproducible reasoning models represent three different facets of the same transformation. AI is no longer a peripheral tool. It is becoming part of the logic, structure, and labor of science itself.


What emerges from this moment is not just a list of tools, it’s about a shift in scientific expectations. Peer review systems need to adapt to AI’s presence, not by locking models out, but by requiring clear records of inputs and contributions. Simulation tools must balance flexibility with staying grounded in the laws of nature, otherwise they risk becoming reflections of data, not representations of theory. And reasoning systems need to be open enough that others can inspect, adjust, and rebuild them. 


In short, the integration of AI into science requires more than technical innovation. It demands architectural forethought about how we represent knowledge, how we assign trust, and how we design systems whose outputs can be tested, revised, and understood.


As the scientific loop closes around AI, we should ask: Are we building tools that expand inquiry or infrastructures that flatten it? Are we gaining new collaborators or just accelerating old assumptions?


The work ahead is not just to scale models, but to ensure they scale the values that make science, science.


This article was created with the assistance of artificial intelligence and thoroughly edited by FirstPrinciples staff and scientific advisors.

 
 
iStock-1357123095.jpg
bottom of page