SR-Scientist: Scientific Equation Discovery With Agentic AI

ICLR 2026 Conference SubmissionAnonymous Authors
Symbolic regressionEquation DiscoveryLarge Language ModelsAgentic AI
Abstract:

Recently, Large Language Models (LLMs) have been applied to scientific equation discovery, leveraging their embedded scientific knowledge for hypothesis generation. However, current methods typically confine LLMs to the role of an equation proposer within search algorithms like genetic programming. In this paper, we present SR-Scientist, a framework that elevates the LLM from a simple equation proposer to an autonomous AI scientist that writes code to analyze data, implements the equation as code, submits it for evaluation, and optimizes the equation based on experimental feedback. Specifically, we wrap the code interpreter into a set of tools for data analysis and equation evaluation. The agent is instructed to optimize the equation by utilizing these tools over a long horizon with minimal human-defined pipelines. Empirical results show that SR-Scientist outperforms baseline methods by an absolute margin of 6% to 35% on datasets covering four science disciplines. Additionally, we demonstrate our method's robustness to noise, the generalization of the discovered equations to out-of-domain data, and their symbolic accuracy. Furthermore, we develop an end-to-end reinforcement learning framework to enhance the agent's capabilities.

Disclaimer
This report is AI-GENERATED using Large Language Models and WisPaper (A scholar search engine). It analyzes academic papers' tasks and contributions against retrieved prior work. While this system identifies POTENTIAL overlaps and novel directions, ITS COVERAGE IS NOT EXHAUSTIVE AND JUDGMENTS ARE APPROXIMATE. These results are intended to assist human reviewers and SHOULD NOT be relied upon as a definitive verdict on novelty.
NOTE that some papers exist in multiple, slightly different versions (e.g., with different titles or URLs). The system may retrieve several versions of the same underlying work. The current automated pipeline does not reliably align or distinguish these cases, so human reviewers will need to disambiguate them manually.
If you have any questions, please contact: mingzhang23@m.fudan.edu.cn

Overview

Overall Novelty Assessment

The paper introduces SR-Scientist, a framework that positions large language models as autonomous agents capable of writing code, analyzing data, and iteratively refining equations based on experimental feedback. This work resides in the LLM-Enhanced Symbolic Regression leaf, which contains five papers total, indicating a moderately populated but still emerging research direction. Unlike sibling papers that primarily use LLMs as equation proposers within traditional search algorithms, SR-Scientist elevates the LLM to a more autonomous role, integrating code interpretation and tool-driven evaluation into the discovery loop.

The broader Symbolic Regression Methods branch encompasses classical genetic programming approaches, reinforcement learning-based methods, and LLM-enhanced techniques. SR-Scientist bridges these areas by combining LLM reasoning with an agent-based workflow, distinguishing it from purely evolutionary methods in the Classical Symbolic Regression leaf and from policy-gradient approaches in the Reinforcement Learning-Based Symbolic Regression leaf. The taxonomy also reveals adjacent directions such as Differential Equation Discovery and Hybrid Neural-Symbolic Methods, which focus on temporal dynamics and neural network integration respectively, rather than the autonomous code-writing paradigm proposed here.

Among thirty candidates examined across three contributions, none were identified as clearly refuting the proposed approach. The SR-Scientist framework contribution examined ten candidates with zero refutable overlaps, as did the reinforcement learning pipeline and tool-driven evaluation system contributions. This suggests that within the limited search scope, the combination of autonomous agent behavior, code-based equation implementation, and iterative optimization through tool use appears distinct from prior LLM-enhanced symbolic regression methods, which typically confine LLMs to hypothesis generation rather than end-to-end scientific workflow orchestration.

The analysis reflects a focused literature search of thirty semantically related papers, not an exhaustive survey of all symbolic regression or LLM-based discovery work. While the statistics indicate no direct prior work overlap within this sample, the relatively small size of the LLM-Enhanced Symbolic Regression leaf and the rapid evolution of LLM-agent research suggest that closely related efforts may exist outside the examined candidate set. The framework's novelty appears strongest in its integration of autonomous coding and tool use, though the reinforcement learning component overlaps conceptually with existing RL-based symbolic regression methods.

Taxonomy

Core-task Taxonomy Papers
50
3
Claimed Contributions
30
Contribution Candidate Papers Compared
0
Refutable Paper

Research Landscape Overview

Core task: scientific equation discovery from observational data. The field is organized around several complementary branches that reflect different problem settings and methodological emphases. Symbolic Regression Methods focus on discovering closed-form algebraic expressions directly from data, often using genetic programming or neural-guided search strategies. Differential Equation Discovery targets systems governed by temporal or spatial dynamics, aiming to recover governing PDEs or ODEs from trajectory observations. Domain-Specific Equation Discovery tailors techniques to particular scientific disciplines such as physics, biology, or climate science, where prior knowledge can guide the search. Causal Discovery and Structure Learning emphasizes identifying not just predictive equations but the underlying causal relationships among variables. Hybrid and Multi-Task Approaches combine symbolic and neural components or tackle multiple discovery objectives simultaneously, while Foundations and Evaluation addresses benchmarking, theoretical guarantees, and reproducibility challenges across the entire landscape. Recent work has seen growing interest in leveraging large language models to enhance symbolic regression, a direction that blends domain knowledge encoded in pre-trained models with classical search heuristics. SR-Scientist[0] exemplifies this trend by integrating LLM-based reasoning into the equation discovery pipeline, positioning itself within the LLM-Enhanced Symbolic Regression cluster alongside LLM-SR[3] and DrSR[2]. While LLM-SR[3] primarily uses language models to propose candidate expressions, SR-Scientist[0] emphasizes a more interactive or scientist-in-the-loop workflow that refines hypotheses iteratively. Nearby efforts such as MLLM-based Discovery of Intrinsic[9] explore multimodal inputs to capture richer contextual cues, and Mimicking the Physicists Eye[30] incorporates visual or geometric reasoning into the discovery process. These LLM-enhanced methods contrast with purely evolutionary or sparsity-driven approaches, raising open questions about the trade-offs between interpretability, computational cost, and the risk of overfitting to patterns seen during pre-training versus genuine physical laws.

Claimed Contributions

SR-SCIENTIST framework for autonomous equation discovery

The authors introduce SR-SCIENTIST, a framework where an LLM agent autonomously discovers scientific equations through long-horizon optimization. The agent uses code interpreters as tools to analyze data and evaluate equations, operating with minimal human-defined pipelines and maintaining an experience buffer to overcome context length limitations.

10 retrieved papers
End-to-end reinforcement learning pipeline for agent capability enhancement

The authors develop a complete RL pipeline including training data construction and reward design that enables the LLM agent to evolve and improve its equation discovery capabilities through self-experience, using GRPO algorithm for optimization.

10 retrieved papers
Tool-driven data analysis and equation evaluation system

The authors design a tool system that wraps code interpreters into two primary tools: a data analyzer for exploring observed data and an equation evaluator for testing hypotheses. This enables the agent to conduct long-horizon optimization through multi-turn interactions without rigid predefined workflows.

10 retrieved papers

Core Task Comparisons

Comparisons with papers in the same taxonomy category

Contribution Analysis

Detailed comparisons for each claimed contribution

Contribution

SR-SCIENTIST framework for autonomous equation discovery

The authors introduce SR-SCIENTIST, a framework where an LLM agent autonomously discovers scientific equations through long-horizon optimization. The agent uses code interpreters as tools to analyze data and evaluate equations, operating with minimal human-defined pipelines and maintaining an experience buffer to overcome context length limitations.

Contribution

End-to-end reinforcement learning pipeline for agent capability enhancement

The authors develop a complete RL pipeline including training data construction and reward design that enables the LLM agent to evolve and improve its equation discovery capabilities through self-experience, using GRPO algorithm for optimization.

Contribution

Tool-driven data analysis and equation evaluation system

The authors design a tool system that wraps code interpreters into two primary tools: a data analyzer for exploring observed data and an equation evaluator for testing hypotheses. This enables the agent to conduct long-horizon optimization through multi-turn interactions without rigid predefined workflows.