Buckets:
Title: Active Symbolic Discovery of Ordinary Differential Equations via Phase Portrait Sketching
URL Source: https://arxiv.org/html/2409.01416
Markdown Content: Back to arXiv
This is experimental HTML to improve accessibility. We invite you to report rendering errors. Use Alt+Y to toggle on accessible reporting links and Alt+Shift+Y to toggle off. Learn more about this project and help improve conversions.
Why HTML? Report Issue Back to Abstract Download PDF Abstract 1Introduction 2Preliminaries 3Methodology 4Related Work 5Experiments 6Conclusion References
HTML conversions sometimes display errors due to content that did not convert correctly from the source. This paper uses the following packages that are not yet supported by the HTML conversion tool. Feedback on these issues are not necessary; they are known and are being worked on.
failed: mdframed
Authors: achieve the best HTML results from your LaTeX submissions by following these best practices.
License: CC BY 4.0 arXiv:2409.01416v2 [cs.LG] null Active Symbolic Discovery of Ordinary Differential Equations via Phase Portrait Sketching Nan Jiang1, Md Nasim2, Yexiang Xue1 1Department of Computer Science, Purdue University, USA 2Department of Computer Science, Cornell University, USA {jiang631, yexiang}@purdue.edu, md.nasim@cornell.edu Abstract
The symbolic discovery of Ordinary Differential Equations (ODEs) from trajectory data plays a pivotal role in AI-driven scientific discovery. Existing symbolic methods predominantly rely on fixed, pre-collected training datasets, which often result in suboptimal performance, as demonstrated in our case study in Figure 1. Drawing inspiration from active learning, we investigate strategies to query informative trajectory data that can enhance the evaluation of predicted ODEs. However, the butterfly effect in dynamical systems reveals that small variations in initial conditions can lead to drastically different trajectories, necessitating the storage of vast quantities of trajectory data using conventional active learning. To address this, we introduce Active Symbolic Discovery of Ordinary Differential Equations via Phase Portrait Sketching (Apps). Instead of directly selecting individual initial conditions, our Apps first identifies an informative region within the phase space and then samples a batch of initial conditions from this region. Compared to traditional active learning methods, Apps mitigates the gap of maintaining a large amount of data. Extensive experiments demonstrate that Apps consistently discovers more accurate ODE expressions than baseline methods using passively collected datasets.
1Introduction
Uncovering the governing principles of physical systems from experimental data is a crucial task in AI-driven scientific discovery [1, 2, 3]. Recent advancements have introduced various methods for uncovering knowledge of dynamical systems in symbolic Ordinary Differential Equation (ODE) form, leveraging techniques such as genetic programming [4], sparse regression [5, 6], Monte Carlo tree search [7], pretrained Transformers [8], and deep reinforcement learning [9].
State-of-the-art approaches discover the symbolic ODEs using a fixed, pre-collected training dataset. However, their performance is often heavily influenced by the quality of the collected data. As illustrated in Figure 1, we find that the best-discovered ODEs from the most recent baseline, that is ODEFormer [10], may fit some test trajectories well, but fit other test trajectories poorly. This observation highlights the need for new methods that actively query informative trajectory data to improve ODE discovery.
Suppose trajectory data can be obtained from a data oracle by specifying the initial conditions. To minimize excessively querying the oracle, a key challenge emerges: given a set of candidate ODEs predicted by a learning method, how can initial conditions within the variable intervals be strategically selected to obtain informative data?
Previous work in the active learning literature typically maintains a large set of data, evaluates their informativeness, and then queries the most informative data points [11, 12]. However, the chaotic nature of dynamical systems complicates the direct application of such methods. The Butterfly effect states that small variations in initial conditions can lead to vastly different outcomes. For instance, as illustrated in Figure 2(c), selecting initial conditions near ( 3 , 0 ) for ๐ 1 can result in trajectories that diverge in opposite directions. Effectively addressing this variability requires densely sampling initial conditions to thoroughly explore the space. Existing active learning-based approaches will be computationally prohibitive and demand significant memory resources, particularly in high-dimensional dynamical systems.
To address these challenges, we propose a novel approach to data querying. We consider selecting a batch of close-neighbor initial conditions instead of individual initial conditions. This process begins by sketching the dynamics in smaller regions, identifying an informative region in the phase space, and then sampling a batch of initial conditions from this region. Figure 2(c) illustrates this idea using phase portraits for three candidate ODEs. Region ๐ข 2 is chosen because the trajectories generated by the candidate ODEs exhibit greater divergence in this region than region ๐ข 1 . Section 3 provides detailed region selection criteria.
Figure 1:The performance of predicted ODE from passively-learned baseline is heavily influenced by the collected training data while our Apps method is not. The dots represent noisy ground-truth trajectory data, and the lines show predicted values of state variables under identical initial conditions. (a, b) Our Apps and the baseline predict accurately for the trajectory starting at ๐ฑ 0
( 0 , 1 ) . (c, d) For the trajectory starting at ๐ฑ 0
( 4 , โ 1 ) , the baseline performs poorly while Apps maintains accuracy.
Thus, we introduce Active Symbolic Discovery of Ordinary Differential Equations via Phase Portrait Sketching (Apps), which consists of two key components: (1) a deep sequential decoder, which guides the search for candidate ODEs by sampling from the defined grammar rules. (2) a data query and evaluation module that actively queries the data using sketched phase portraits and evaluates the candidate ODE. In experiments, we evaluate Apps against several popular baselines on two large-scale ODE datasets. 1) Apps achieves the lowest median NMSE (in Table 1 and Table 2) across multiple datasets under noiseless and noisy settings. 2) Compared to other active learning strategies, Apps is more time efficient in benchmark datasets (in Table 3). 1.
2Preliminaries
Ordinary Differential Equations (ODEs) describe the evolution of dynamical systems in continuous time. Let vector ๐ฑ โข ( ๐ก )
( ๐ฅ 1 โข ( ๐ก ) , โฆ , ๐ฅ ๐ โข ( ๐ก ) ) โ โ ๐ be the state variables of the system of time ๐ก . The temporal evolution of the system is governed by the time derivatives of the state variables, denoted as ๐ โข ๐ฅ ๐ ๐ โข ๐ก . The general form of the ODE is written as:
๐ โข ๐ฅ ๐ ๐ โข ๐ก
๐ ๐ โข ( ๐ฑ โข ( ๐ก ) , ๐ ) , for โข ๐
1 , โฆ , ๐ ,
where ๐ ๐ can be a linear or nonlinear function of the state variables ๐ฑ and coefficients ๐ . The ODE is noted as a tuple ( ๐ 1 , ๐ 2 , โฆ , ๐ ๐ ) for simplicity in this paper. Function ๐ ๐ is symbolically expressed using a subset of input variables in ๐ฑ and coefficients in ๐ , connected by mathematical operators such as addition and cosine functions. For example, we use ( 10 โข sin โก ( ๐ฅ 2 ) , 4 โข cos โก ( ๐ฅ 1 + 2 ) ) to represent the ODE ๐ โข ๐ฅ 1 ๐ โข ๐ก
10 โข sin โก ( ๐ฅ 2 ) , ๐ โข ๐ฅ 2 ๐ โข ๐ก
4 โข cos โก ( ๐ฅ 1 + 2 ) .
Given an initial condition ๐ฑ 0 , the solution to the ODE is a trajectory of state variables ( ๐ฑ 0 , ๐ฑ โข ( ๐ก 1 ) , โฆ , ๐ฑ โข ( ๐ก ๐ ) ) observed at discrete time points ( ๐ก 1 , โฆ , ๐ก ๐ ) , possibly with noise. The trajectory is noted as ๐ for simplicity.
Phase Portrait is a qualitative analysis tool for studying the behavior of dynamical systems [13]. Phase portraits are plotted using the state variables ๐ฑ and their time derivatives ( ๐ 1 , โฆ , ๐ ๐ ) . A curve in the phase portrait is a short trajectory of the system over time from a given initial condition. The arrow on the curve indicates the direction of change. By examining these curves, we can infer key properties of the system, such as stability, equilibrium points, and periodic behavior. Figure 2(c) shows phase portraits for three different ODEs. These portraits are generated by sampling random initial conditions within the variable intervals and evolving the system for a short time.
Symbolic Discovery of Ordinary Differential Equations seeks to uncover the symbolic form of an ODE that best fits a dataset of observed trajectories. According to Gec et al. [14] and Sun et al. [7], we are given a dataset of collected trajectories ๐ท
{ ๐ 1 , โฆ , ๐ ๐ } and a set of mathematical operators { + , โ , ร , รท , sin โก โฆ } . Denote ๐ โข ( ๐ฑ โข ( ๐ก ) , ๐ ) as a candidate ODE, where ๐ indicates the coefficients. The objective is to predict the symbolic form of the ODE that minimizes the distance between the predicted and observed trajectories, which is formalized as an optimization problem:
arg
โก
min
๐
โ
ฮ
โก
1
|
๐ท
|
โ
๐
โ
๐ท
โ
๐
1 ๐ โ โข ( ๐ฑ โข ( ๐ก ๐ ) , ๐ฑ ^ โข ( ๐ก ๐ ) ) ,
(1)
where
๐ฑ ^ โข ( ๐ก ๐ )
๐ฑ 0 + โซ 0 ๐ก ๐ ๐ โข ( ๐ฑ โข ( ๐ก ) , ๐ ) โข ๐ ๐ก .
ฮ is the set of all possible ODEs, trajectory ๐ := ( ๐ฑ 0 , ๐ฑ ( ๐ก 1 ) , โฆ , ๐ฑ ( ๐ก ๐ ) ) , ๐ฑ โข ( ๐ก ) is the ground-truth observations of the state variable. Trajectory ( ๐ฑ 0 , ๐ฑ ^ ( ๐ก 1 ) , โฆ , ๐ฑ ^ ( ๐ก ๐ ) ) is the predicted state variables according to the candidate ODE ๐ . The predicted trajectory ( ๐ฑ 0 , ๐ฑ ^ โข ( ๐ก 1 ) , โฆ , ๐ฑ ^ โข ( ๐ก ๐ ) ) is obtained by numerically integrating the ODE from the given initial state ๐ฑ 0 to the final time ๐ก ๐ . The loss function โ computes the summarized distance between the predicted and ground-truth trajectories at each time step. A typical loss is the Normalized Mean Squared Error (NMSE, defined in Appendix D). Except for the above formulation, prior works in symbolic regression use the approximated time derivative as the label to discover each expression ๐ ๐ separately, which is known as gradient matching. We leave the discussion to Related Work.
Figure 2:The pipeline of Apps for symbolic discovery of ODEs consists of 3 steps: (a) ODEs are sampled from the sequential decoder by iteratively sampling grammar rules. The predicted rule at each step serves as input for the decoder in the subsequent step. (b) The sampled sequence of grammar rules is converted into a valid ODE with ๐
2 variables. Each rule expands the first non-terminal symbol, with the expanded parts highlighted in blue colors for clarity. (c) The phase portrait for the predicted ODEs (e.g., ๐ 1 , ๐ 2 , ๐ 3 ) is sketched, and regions with high informativeness, such as ๐ข 2 , are identified to query the new trajectory data. In region ๐ข 2 , ๐ 1 exhibits a saddle point, ๐ 2 moves downward, and ๐ 3 moves upward. In contrast, in region ๐ข 1 , all trajectories move from right to left. Differentiating the predicted expressions is easier in region ๐ข 2 than in region ๐ข 1 .
Recent research explored deep reinforcement learning to discover the governing equations from data [15, 16, 17]. In these approaches, each expression is represented as a binary tree, with interior nodes corresponding to mathematical operators and leaf nodes to variables or constants. An ODE with ๐ variables is represented by ๐ trees. The key idea is to frame the search for different ODEs as a sequential decision-making process based on the preorder traversal sequence of expression trees. A high reward is assigned to candidates which fit the data well. The search is guided by a deep sequential decoder, often based on RNN, LSTM, or decoder-only Transformer, that learns the optimal probability distribution for selecting the next node in the expression tree at each step. The parameters of the decoder are trained with the policy gradient algorithm.
3Methodology 3.1Motivation
For the task of symbolic discovery of ODEs, we observe that existing methods frequently overfit the training data. This issue is illustrated in Figure 1 using ODEFormer [10], a recent baseline designed to learn ODEs from a fixed training dataset. In the example, the best-predicted ODE is given by ๐
( 1.04 โข ๐ฅ 2 , โ 0.02 โ 0.77 โข ๐ฅ 1 ) . We evaluate ๐ on noisy test trajectories (depicted as blue dots) with two distinct initial conditions. While ๐ closely aligns with the trajectory originating at ๐ฑ 0
( 0 , 1 ) , as shown by the green curve, it produces substantial errors for a trajectory starting at ๐ฑ 0
( 4 , โ 1 ) , where the predicted curve deviates significantly from the ground truth.
This observation motivates us to actively identify informative trajectory data to better differentiate candidate expressions during the learning process. Each trajectory is generated by querying the data oracle with a specified initial condition ๐ฑ 0 . An initial condition is deemed informative if the resulting trajectory for different candidate ODEs diverges significantly. The key challenge lies in selecting such informative initial conditions from the variable intervals for a given set of candidate ODEs.
For addressing this issue, a common approach in active learning [11] is to maintain a large set of potential initial conditions, evaluate their informativeness, and query the most informative points. However, the butterfly effect in chaos theory [18] suggests existing works in active learning are not directly applicable. The chaotic behavior states small changes in initial conditions can lead to drastically different outcomes in dynamical systems. For example, as shown in Figure 2(c), selecting points near ( 3 , 0 ) (inside the red region ๐ข 2 ) for ๐ 1 can lead to trajectories diverging either towards the top right or the bottom left. Such chaotic behavior necessitates the existing active learning methods to maintain a large set of initial conditions to adequately cover the domain, which becomes infeasible for high-dimensional dynamical systems.
To mitigate this issue, we consider selecting a beam of near-neighbor points rather than individual points. We propose first to select a highly informative region and sample a batch of initial conditions within that region. In this research, the region is represented as an ๐ -dimensional cube of fixed width. A region is regarded as informative if the majority of sampled initial conditions within it yield informative trajectories for the given candidate ODEs.
Figure 2(c) illustrates our region-based approach using the phase portraits of three candidate ODEs: ๐ 1 , ๐ 2 , and ๐ 3 . Each curve in the phase portrait represents a short trajectory, with its starting point and direction indicating the initial conditions and the direction of evolution over time. A closer look reveals significant differences in dynamics within region ๐ข 2 across the ODEs. While the curves in region ๐ข 1
[ โ 2 , 0 ] ร [ โ 2 , 0 ] consistently move from the bottom right to the top left in all phase portraits, the trajectories in region ๐ข 2
[ 2 , 4 ] ร [ โ 1 , 1 ] exhibit drastically different behaviors. This indicates that trajectories originating from region ๐ข 2 are more divergent and thus more informative.
Main Procedure. The proposed Apps, illustrated in Figure 2, comprises two key components: (1) Deep Sequential Decoder. This module predicts candidate ODEs by sampling sequences of grammar rules defined for symbolic ODE representation. (2) Data Sampling Module. Using the proposed phase portrait sketching, this module selects a batch of informative ground-truth data points.
Throughout the training process, the reward for the predicted ODEs is computed using the queried data, and the decoder parameters are updated via policy gradient estimation. Among all sampled candidates, Apps selects the ODE with the smallest loss value (as defined in Equation 1) as the final prediction.
Connection to Existing Approaches. Like dโAscoli et al. [10], Apps employs a Transformer-based decoder. However, unlike dโAscoli et al. [10], which learns from fixed data, Apps actively queries new data. The learning objective of Apps is inspired by Petersen et al. [15], where both approaches guide the search for the optimal equation as a decision-making process over a sequence of tokens.
Existing active learning methods, particularly in symbolic regression, have largely overlooked the chaotic behaviors inherent in dynamical systems. For instance, Jin et al. [19] proposed a separate generative model for sampling informative data, assuming that input data within a small region should exhibit minimal output divergence. However, this assumption fails to hold in the context of dynamical systems. Additionally, Haut et al. [20] formulated an optimization problem based on the Query-By-Committee (QBC) method in active learning, to find those informative initial conditions. But the optimization needs to maintain a large set of data points, to account for the chaotic behaviors. The rest of the discussion is provided in the Related Work.
3.2The Learning Pipeline
Data Assumption. Our method relies on the assumption that we can query a data oracle ๐ช by specifying the initial conditions ๐ฑ 0 and discrete times ๐
( ๐ก 1 , โฆ , ๐ก ๐ ) . The oracle executes ๐ช โข ( ๐ฑ 0 , ๐ ) and returns a (noisy) observation of the trajectory at the specified discrete times ๐ . In science, this data query process is achieved by conducting real-world experiments with specified configurations. Recent work [21, 22, 23] also highlight the importance of having the oracle that can actively query data points, rather than learning from a fixed dataset.
Expression Representation. To enable the sequential decoder to predict an ODE by generating a sequence step-by-step, we extend the context-free grammar to represent an ODE as a sequence of grammar rules [24, 14, 7]. The grammar is defined by the tuple โจ ๐ , ฮฃ , ๐ , ๐ โฉ , where ๐ is a set of non-terminal symbols, ฮฃ is a set of terminal symbols, ๐ is a set of production rules and ๐ โ ๐ is the start symbol.
More specifically, each component of the grammar is: 1) For the non-terminal symbols, we use ๐ด to represent a sub-expression for ๐ โข ๐ฅ 1 ๐ โข ๐ก and ๐ต to represent a sub-expression for ๐ โข ๐ฅ 2 / โข ๐ โข ๐ก . For dynamical systems with ๐ variables, we use ๐ distinct non-terminal symbols. 2) The terminal symbols include the input variables and constants { ๐ฅ 1 , โฆ , ๐ฅ ๐ , ๐๐๐๐๐ } . 3) The production rules correspond to mathematical operations. For example, the addition operation is represented as ๐ด โ ( ๐ด + ๐ด ) , where the rule replaces the left-hand symbol with the right-hand side. 4) The start symbol is redefined as โ ๐ โ ๐ด , ๐ต โ, where the comma notation indicates that ๐ด and ๐ต represent two separate equations in a two-variable dynamical system. Similarly, there will be ๐ non-terminal symbols connected by ๐ โ 1 comma for ๐ -dimensional dynamical system.
Starting from the start symbol, different symbolic ODEs are constructed by applying the grammar rules in various sequences. An ODE is valid if it only consists of terminal symbols; otherwise, it is invalid. Figure 2(b) provides an example of constructing the ODE ๐ โข ๐ฅ 1 ๐ โข ๐ก
๐ฅ 2 , ๐ โข ๐ฅ 2 ๐ โข ๐ก
โ 0.9 โข sin โก ( ๐ฅ 1 ) from the start symbol ๐ โ ๐ด , ๐ต using a sequence of grammar rules. The replaced parts are color highlighted. Initially, the multiplication rule ๐ต โ ๐ต ร ๐ต is applied, replacing the symbol ๐ต in ๐ 2
๐ต with ๐ต ร ๐ต , resulting in ๐ โ ๐ด , ๐ต ร ๐ต . Next, the rule ๐ด โ ๐ฅ 2 is applied, yielding ๐ โ ๐ฅ 2 , ๐ต ร ๐ต . Iteratively applying the rules, we eventually obtain ๐ โ ๐ฅ 2 , ๐ 1 ร sin โก ( ๐ฅ 1 ) , which corresponds to one candidate ODE ๐
( ๐ฅ 2 , ๐ 1 โข sin โก ( ๐ฅ 1 ) ) . The coefficient ๐ 1
โ 0.9 is obtained when fitting to the trajectory data. The procedure of coefficient fitting is described in Appendix C โImplementation of Appsโ section.
Sampling ODEs from Decoder. The proposed Apps is built on top of a sequential decoder, which generates different ODEs as a sequential decision-making process. The decoder can be RNN, LSTM, or the decoder-only Transformer. The input and output vocabulary is the set of allowed rules covering input variables, coefficients, and mathematical operators. Predicting ODEs involves using the decoder to sample a sequence of grammar rules, where each sequence corresponds to a candidate ODE using previously defined grammar. The objective of Apps is to maximize the probability of sampling those ODEs that fit the data well. This is achieved through the REINFORCE objective, where the objective computes the expected reward of ODE to the data. In our formulation, the reward is evaluated on selected data by the phase portrait sketching module.
As shown in Figure 2(a), the decoder receives the start symbol ๐ 0
โ ๐ โ ๐ด , ๐ต โ and outputs a categorical distribution ๐ ๐ โข ( ๐ 1 | ๐ 0 ) over rules in the output vocabulary. This distribution represents the probabilities of possible next rules in the partially completed expression. One token is drawn from this distribution, ๐ 1 โผ ๐ โข ( ๐ 1 | ๐ 0 ) , which serves as the prediction for the second rule and is used as the input for the next step. At ๐ก -th step, the predicted output from the previous step ๐ ๐ก is used as the input for the current step. The decoder draws rule ๐ ๐ก + 1 from the probability distribution ๐ ๐ก + 1 โผ ๐ ๐ โข ( ๐ ๐ก + 1 | ๐ 0 , โฆ , ๐ ๐ก ) . This process iterates until maximum steps are reached, with a probability of ๐ ๐ โข ( ๐ )
โ ๐
1 ๐ โ 1 ๐ ๐ โข ( ๐ ๐ | ๐ 1 , โฆ , ๐ ๐ โ 1 ) . The sampled sequence is converted into an expression following the definition previously described in โExpression Representationโ.
Active Query Data with Phase Portrait Sketching. To evaluate the goodness-of-fit of generated ODEs from the decoder and differentiate which one is better, we propose comparing the phase portrait of predicted ODEs. We sketch the phase portrait using collections of short trajectories, all starting from the same initial conditions and sharing the same time sequence.
Following our discussion in the โMotivationโ section, a region is considered informative for distinguishing between two candidate ODEs if their sketched phase portraits differ. To identify such regions, we randomly sample several and sketch the phase portraits for all candidate ODEs within each. The most informative region is then selected, and we query the data oracle (noted as ๐ช ) for the ground-truth trajectory in that region.
Formally, assume we are given ๐ ODEs, { ๐ 1 , โฆ , ๐ ๐ } , and ๐พ randomly selected regions, { ๐ข 1 , โฆ , ๐ข ๐พ } . Each region ๐ข ๐ is a Cartesian product of ๐ intervals, expressed as ๐ข ๐
[ ๐ 1 , ๐ 1 ] ร โฏ ร [ ๐ ๐ , ๐ ๐ ] . To sketch the dynamics of candidates in the region ๐ข ๐ , we uniformly sample ๐ฟ points in ๐ข ๐ , ๐ฑ 1 , โฆ , ๐ฑ ๐ฟ , as initial conditions. For region ๐ข ๐ , the trajectory ๐ ๐ , ๐ , ๐
( ๐ฑ ๐ , ๐ฑ ^ โข ( ๐ก 1 ) , โฆ , ๐ฑ ^ โข ( ๐ก ๐ ) ) is generated by the expression ๐ ๐ , starting from the ๐ -th initial condition ๐ฑ ๐ and evolving over time according to the numerical integration ๐ฑ ^ โข ( ๐ก ๐ )
๐ฑ ๐ + โซ 0 ๐ก ๐ ๐ ๐ โข ( ๐ฑ โข ( ๐ก ) , ๐ ) โข d ๐ก for ๐ก ๐ โ { ๐ก 1 , โฆ , ๐ก ๐ } . The resulting ๐ฟ short trajectories form a sketched phase portrait for ODE ๐ ๐ in the region ๐ข ๐ .
Two expressions, ๐ ๐ and ๐ ๐ โฒ , have similar sketches in region ๐ข ๐ if their corresponding trajectories, starting from the same initial condition, are close. Specifically, this occurs when โ ๐
1 ๐ฟ โ ๐ ๐ , ๐ , ๐ โ ๐ ๐ โฒ , ๐ , ๐ โ โ 0 . We define the pairwise informative score between ๐ ๐ and ๐ ๐ โฒ in region ๐ข ๐ as:
๐ธ๐ต โข ( ๐ ๐ , ๐ ๐ โฒ , ๐ข ๐ )
1 ๐ฟ โข โ ๐
1 ๐ฟ โ ๐ ๐ , ๐ , ๐ โ ๐ ๐ โฒ , ๐ , ๐ โ 2 2
(2)
The total informative score for a region (denoted as ๐ธ๐ต โข ( ๐ข ๐ ) ) is the sum of the pairwise informative scores for every pair of candidate ODEs. The informative score for region ๐ข ๐ is:
๐ธ๐ต โข ( ๐ข ๐ )
1 ๐ โข โ ๐
1 ๐ โ ๐ โฒ
๐ + 1 ๐ ๐ธ๐ต โข ( ๐ ๐ , ๐ ๐ โฒ , ๐ข ๐ )
(3)
We select the region with the highest informative score, denoted ๐ข โ โ arg โก max ๐
1 ๐พ โก ๐ธ๐ต โข ( ๐ข ๐ ) . A batch of ๐ initial conditions, { ๐ฑ 1 , โฆ , ๐ฑ ๐ } , is then sampled from region ๐ข โ , and the data oracle ๐ช โข ( ๐ฑ ๐ , ๐ ) is queried with the given initial conditions. The obtained ground-truth trajectories are used to compute the reward function for the objective, which in turn updates the modelโs parameters. In practice, the relative size of the regions and the number of sampled regions are set as hyper-parameters in the experiments.
Policy Gradient-based Training. The REINFORCE objective that maximizes the expected reward is
๐ฝ โข ( ๐ ) := ๐ผ ๐ โผ ๐ ๐ โข ( ๐ ) โข [ ๐๐๐ ๐๐๐ โข ( ๐ ) ]
where ๐ ๐ โข ( ๐ ) is the probability of sampling sequence ๐ and ๐ represents the parameters of the decoder. Following the REINFORCE policy gradient algorithm [25], the gradient w.r.t. the objective โ ๐ ๐ฝ โข ( ๐ ) is estimated by the empirical average over the samples from the probability distribution ๐ ๐ โข ( ๐ ) . We first sample ๐ sequences ( ๐ 1 , โฆ , ๐ ๐ ) , and an unbiased estimation of the gradient of the objective is computed as:
โ ๐ ๐ฝ โข ( ๐ ) โ 1 ๐ โข โ ๐
1 ๐ ๐๐๐ ๐๐๐ โข ( ๐ ๐ ) โข โ ๐ log โก ๐ ๐ โข ( ๐ ๐ )
The parameters of the decoder are updated using the estimated policy gradient value. This update process increases the probability of generating high goodness-of-fit ODEs. Detailed derivations are presented in Appendix C.
4Related Work
AI-driven Scientific Discovery. Artificial intelligence has increasingly been employed to accelerate discoveries in learning ordinary and partial differential equations directly from data [5, 3, 2, 26, 27, 28, 29, 30, 31, 32].
Strogatz dataset (
๐ 2
0 , ๐ผ
0 ) ODEbase dataset ( ๐ 2
0 , ๐ผ
0
)
๐
1
๐
2
๐
3
๐
4
๐
2
๐
3
๐
4
๐
5
SPL 0.787
0.892
1.921
2.865 0.867 2.17 4.75 13.16 E2ETransformer 6.47 โข ๐ธ โ 4
1.620
๐ . ๐ .
๐ . ๐ . 0.757 ๐ . ๐ .
๐ . ๐ .
๐ . ๐ .
ProGED 0.129
0.666
2.68
3.856 0.317 2.134 ๐ . ๐ .
๐ . ๐ .
SINDy 1.90 โข ๐ธ โ 4 0.217 1.539
4.810 0.521 2.112 8.334 52.12 ODEFormer 0.0303
0.9261
1.033
1.010 0.213 0.245 1.213
3.148
Apps (ours)
2.06
โข
๐
โ
๐
0.2912
1.011 0.521
0.1318
0.1306
1.046
3.054
Table 1:On the noiseless datasets with regular time sequence (
๐
2
0 , ๐ผ
0 ), Median NMSE is reported over the best-predicted expression found by all the algorithms. Our Apps method can discover the governing expressions with smaller NMSE values than baselines, under the noiseless setting. T.O. means termination with a 24-hour limit.
Symbolic Regression for ODEs. Symbolic regression, traditionally used to identify algebraic equations between input variables and output labels, has been extended to discover ODEs. A key ingredient is gradient matching, which approximates labels for symbolic regression by using finite differences of consecutive states along a trajectory [7, 33, 8, 14]. Recent methods, such as SINDy and its extensions [5, 34], leverage sparse regression techniques to directly learn the structure of ODEs and PDEs from data. They perform particularly well with trajectory data sampled at small, regular time intervals, where the approximations closely align with true derivatives.
Neural Networks Learns Implicit ODEs. This research direction involves learning ODE implicitly. Early work employed Gaussian Processes to model ODEs [35]. Neural ODEs further advanced the field by parameterizing ODEs with neural networks, enabling training through backpropagation via ODE solvers [32]. Physics-informed neural networks integrate physical knowledge, such as conservation laws, into the modeling process [29]. Meanwhile, Fourier neural operators use neural networks to learn the functional representation [36].
Active Learning aims to query informative unlabeled data to accelerate convergence with fewer samples [37, 38, 39, 40]. In symbolic regression, query-by-committee strategies have been explored to actively query data for discovering algebraic equations [41, 23]. For example, Jin et al. [19] proposed a method that learns uncertainty distributions using neural networks and queries data with high uncertainty. However, all these methods largely overlooked the chaotic behaviors inherent in dynamical systems.
5Experiments
This section shows our Apps can find ODEs with the smallest errors (Normalized MSE) among all competing approaches, under noiseless, noisy, and irregular time settings (see Table 1 and Table 2). Compared to the baselines, our Apps data query strategy requires fewer data and attains a better ranking of the TopK candidate ODEs (see Table 3).
5.1Experimental Settings
Datasets. We consider 2 datasets of multivariate variables, including (1) Strogatz dataset [10] of 80 instances, collected from the Strogatz textbook [13]. It is formalized as a benchmark dataset by [10]. (2) ODEBase dataset [42] of 114 instances, containing equations from chemistry and biology. Each dataset is further partitioned by the number of variables contained in the ODE.
We consider 3 different conditions: (1) regular time noiseless condition, (2) regular time noisy condition, and (3) irregular time condition. In the noiseless setting, the obtained data is exactly the evaluation of the ground-truth expression. In the noisy setting, the obtained data is further perturbed by Gaussian noise. We add multiplicative noise by replacing each ๐ฑ โข ( ๐ก ๐ ) with ( 1 + ๐ ) โข ๐ฑ โข ( ๐ก ๐ ) , and ๐ is sampled from a zero mean multivariate Gaussian distribution with diagonal variances ๐๐๐๐ โข ( ๐ 2 , โฆ , ๐ 2 ) . The noise rate is determined by ๐ 2 . For both noiseless and noisy settings, the data points are sampled at regular time intervals. In the irregular time setting, we first generate the regular time sequence and drop a fraction with probability ๐ผ . The rate of time irregularity is determined by ๐ผ .
Noisy Strogatz datasets (
๐ 2
0.01 , ๐ผ
0 ) Irregular Strogatz dataset ( ๐ 2
0 , ๐ผ
0.1
)
๐
1
๐
2
๐
3
๐
4
๐
1
๐
2
๐
3
๐
4
SPL 0.938
1.019
2.915
3.068
0.127
0.526
3.196
4.193
SINDy 6.4 โข ๐ธ โ 3 4.152 2.498 5.21 6.66 โข ๐ธ โ 4 0.472 0.827
4.163
ProGED 0.121
0.658
3.673
3.856
0.134
0.769
2.766
4.181
ODEFormer 0.139
0.621
2.392
0.812
0.031
1.036 1.51 1.011 Apps (ours) 7.75E-4 0.369 1.381 0.657 1.06E-6 0.215 1.012 0.947 Table 2:On the Strogatz dataset, the Median NMSE is reported over the best-predicted expression found by all the algorithms under noisy or irregular time sequence settings.
Baselines. We consider a line of recent works for symbolic equation discovery as our baselines. The methods using passive data query strategy are as follows: (1) SINDy [5], (2) ODEFormer [10], (3) Symbolic Physics Learner (SPL) [7], (4) Probabilistic grammar for equation discovery (ProGED) [14], (5) end-to-end Transformer (E2ETransformer) [43].
Evaluation. For evaluating all the methods, we considered 3 different metrics: (1) goodness-of-fit using NMSE, (2) empirical running time of data querying step, and (3) ranking-based distance. The goodness-of-fit using the NMSE indicates how well the learning algorithms perform in discovering symbolic expressions. Given the best-predicted expression by each algorithm, we evaluate the goodness-of-fit on a larger testing set with longer time steps and a larger batch size of data. The median (50%) of the NMSE is reported in the benchmark table. The full quantiles ( 25 % , 50 % , 75 % ) of the NMSE are further provided. The remaining details of the experiment settings are in Appendix D.
5.2Experimental Analysis
Goodness-of-fit Benchmark. We summarize our Apps on several challenging multivariate datasets with noiseless data in Table 1. It shows our Apps attains the smallest median NMSE values on all datasets, against a line of current popular baselines. The performance of SPL and E2Etransformer drops greatly on irregular time sequences because the approximated time derivative becomes inaccurate when missing the intermediate sequence. Our Apps does not suffer from that because it outputs the predicted trajectory and does not need to approximate the time derivative. Another reason is the decoder with massive parameters can better adapt to actively collected datasets.
Figure 3:On the selected data (Strogatz dataset with ๐
1 ), quartiles of NMSE and ๐ 2 scores of the learning algorithms.
Noisy and Irregular Time Settings. We examine the performance of predicting trajectories in the presence of noise and irregular time sequences in Table 2. The ground-truth trajectory is subject to Gaussian noise with zero mean and ๐ 2
0.05 , and an irregularly sampled sequence where 50 % of evenly spaced points are uniformly dropped. The predicted trajectory by each algorithm is compared against the ground truth, utilizing identical initial conditions. Our Apps still attains a relatively smaller NMSE against baselines under the two settings.
Quantiles of Evaluation Metrics. We further report the quantiles of the NMSE metric in Figure 3 to assist the result in Table 1(a). Note that we cut off the negative values as zero when demonstrating ๐ 2 score. The two box plots in Figure 3 show the proposed Apps is consistently better than the baselines in terms of the full quantiles ( 25 % , 50 % , 75 % ) of the NMSE metric.
Benchmark with other Active Strategies. Two baseline methods using active learning strategy are: (1) query-by-committee (QbC) proposed in [41, 23]. (2) Core-Set [39] proposes to sample diverse data. These methods were originally proposed with different neural networks, thus we evaluate these different active learning methods using the same decoder in our Apps. Current active learning methods are not directly available for evaluation in our problem setting (in Equation 1), so we re-implement these query strategies with the new problem setting.
The running time of the data querying step measures the efficiency of every active learning algorithm for this task. The ranking-based distance indicates if the ranking of many candidate expressions is exactly the same as evaluated on full data. If the predicted ODEs are ranked in the same order as the full data, then the ranking-based distance (Kendall tau score) will be close to zero.
In Table 3, given a set of 20 predicted ODEs, we compare the TopK ranking (i.e., top 3) of predicted ODEs by each active learning strategy is the same as using full data. We find both our phase portrait and QbC rank those predicted ODEs in proper ranking order. Our Apps takes the least memory to locate the most informative region and is also time efficient because we only pick one region among all the available regions. The QbC takes much more time because it finds every initial condition as an optimization problem over the input variables, which is solved by a separate gradient-based optimizer. CoreSet first runs a clustering algorithm over the ground-truth data and then samples a diverse set of initial conditions from each cluster. So the memory usage of Coreset is mainly determined by the first clustering step.
Ranking-based Running Peak
distance (
โ ) Time ( โ ) Memory ( โ ) Apps (ours) 0.08
5.2 sec 3.76 MB QbC 0.13
13.4 sec 51 MB CoreSet 0.22
4.3 sec 2.74 GB Table 3:Ranking comparison with different active learning strategies. Apps shows a smaller ranking-based distance than other strategies, which is better for ranking those best-predicted expressions. Also Apps takes less memory consumption and less computational time because the sketching step itself is lightweight. 6Conclusion
In this paper, we introduced Apps, a novel approach for discovering ODEs from trajectory data. By actively reasoning about the most informative regions within the phase portrait of candidate ODEs, Apps overcomes the limitations of passively learned methods that rely on pre-collected datasets. Our approach also reduces the need for extensive data collection while still yielding highly accurate and generalizable ODE models. The experimental results demonstrate that Apps consistently outperforms baseline methods, achieving the lowest median NMSE across various datasets under both noiseless and noisy conditions.
Acknowledgments
We thank all the reviewers for their constructive comments. This research was supported by NSF grant CCF-1918327, NSF Career Award IIS-2339844, DOE โ Fusion Energy Science grant: DE-SC0024583, and AI Climate: NSF and USDA-NIFA and Cornell University AI for Science Institute.
References Schmidt and Lipson [2009] โ Michael Schmidt and Hod Lipson.Distilling free-form natural laws from experimental data.Science, 324(5923):81โ85, 2009. Zhang and Lin [2018] โ Sheng Zhang and Guang Lin.Robust data-driven discovery of governing physical laws with error bars.Proc Math Phys Eng Sci., 474(2217), 2018. Wu and Tegmark [2019] โ Tailin Wu and Max Tegmark.Toward an artificial intelligence physicist for unsupervised learning.Phys. Rev. E, 100:033311, Sep 2019. He et al. [2022] โ Baihe He, Qiang Lu, Qingyun Yang, Jake Luo, and Zhiguang Wang.Taylor genetic programming for symbolic regression.In GECCO, pages 946โ954, 2022. Brunton et al. [2016] โ Steven L. Brunton, Joshua L. Proctor, and J. Nathan Kutz.Discovering governing equations from data by sparse identification of nonlinear dynamical systems.PNAS, 113(15):3932โ3937, 2016. Fasel et al. [2022] โ Urban Fasel, J Nathan Kutz, Bingni W Brunton, and Steven L Brunton.Ensemble-sindy: Robust sparse model discovery in the low-data, high-noise limit, with active learning and control.Proceedings of the Royal Society A, 478(2260):20210904, 2022. Sun et al. [2023] โ Fangzheng Sun, Yang Liu, Jian-Xun Wang, and Hao Sun.Symbolic physics learner: Discovering governing equations via monte carlo tree search.In ICLR. OpenReview.net, 2023. Qian et al. [2022] โ Zhaozhi Qian, Krzysztof Kacprzyk, and Mihaela van der Schaar.D-CODE: discovering closed-form odes from observed trajectories.In ICLR. OpenReview.net, 2022. Jiang et al. [2024] โ Nan Jiang, Md. Nasim, and Yexiang Xue.Vertical symbolic regression via deep policy gradient.In IJCAI, pages 5891โ5899. ijcai.org, 2024. dโAscoli et al. [2024] โ Stรฉphane dโAscoli, Sรถren Becker, Alexander Mathis, Philippe Schwaller, and Niki Kilbertus.Odeformer: Symbolic regression of dynamical systems with transformers.In ICLR. OpenReview.net, 2024. Golovin et al. [2010] โ Daniel Golovin, Andreas Krause, and Debajyoti Ray.Near-optimal bayesian active learning with noisy observations.In NIPS, pages 766โ774. Curran Associates, Inc., 2010. Medina and White [2023] โ Jorge Medina and Andrew D White.Active learning in symbolic regression performance with physical constraints.arXiv preprint arXiv:2305.10379, 2023. Strogatz [2018] โ Steven H Strogatz.Nonlinear dynamics and chaos: with applications to physics, biology, chemistry, and engineering.CRC press, 2018.ISBN 9780429961113. Gec et al. [2022] โ Bostjan Gec, Nina Omejc, Jure Brence, Saso Dzeroski, and Ljupco Todorovski.Discovery of differential equations using probabilistic grammars.In DS, volume 13601, pages 22โ31. Springer, 2022. Petersen et al. [2021] โ Brenden K. Petersen, Mikel Landajuela, T. Nathan Mundhenk, Clรกudio Prata Santiago, Sookyung Kim, and Joanne Taery Kim.Deep symbolic regression: Recovering mathematical expressions from data via risk-seeking policy gradients.In ICLR. OpenReview.net, 2021. Abolafia et al. [2018] โ Daniel A. Abolafia, Mohammad Norouzi, and Quoc V. Le.Neural program synthesis with priority queue training.CoRR, abs/1801.03526, 2018. Mundhenk et al. [2021] โ T. Nathan Mundhenk, Mikel Landajuela, Ruben Glatt, Clรกudio P. Santiago, Daniel M. Faissol, and Brenden K. Petersen.Symbolic regression via deep reinforcement learning enhanced genetic programming seeding.In NeurIPS, pages 24912โ24923, 2021. Lorenz [1963] โ E Lorenz.Deterministic nonperiodic flow in journal of the atmospheric science.1963. Jin et al. [2023] โ Pengwei Jin, Di Huang, Rui Zhang, Xing Hu, Ziyuan Nan, Zidong Du, Qi Guo, and Yunji Chen.Online symbolic regression with informative query.In AAAI, pages 5122โ5130. AAAI Press, 2023. Haut et al. [2024] โ Nathan Haut, Wolfgang Banzhaf, and Bill Punch.Active learning in genetic programming: Guiding efficient data collection for symbolic regression.IEEE Transactions on Evolutionary Computation, pages 1โ13, 2024.doi: 10.1109/TEVC.2024.3471341. Chen and Xue [2022] โ Qi Chen and Bing Xue.Generalisation in genetic programming for symbolic regression: Challenges and future directions.In Women in Computational Intelligence: Key Advances and Perspectives on Emerging Topics, pages 281โ302. Springer, 2022. Keren et al. [2023] โ Liron Simon Keren, Alex Liberzon, and Teddy Lazebnik.A computational framework for physics-informed symbolic regression with straightforward integration of domain knowledge.Scientific Reports, 13(1):1249, 2023. Haut et al. [2023] โ Nathan Haut, Bill Punch, and Wolfgang Banzhaf.Active learning informs symbolic regression model development in genetic programming.In GECCO Companion, pages 587โ590, 2023. Todorovski and Dzeroski [1997] โ Ljupco Todorovski and Saso Dzeroski.Declarative bias in equation discovery.In ICML, pages 376โ384. Morgan Kaufmann, 1997. Williams [1992] โ Ronald J. Williams.Simple statistical gradient-following algorithms for connectionist reinforcement learning.Mach. Learn., 8:229โ256, 1992. Iten et al. [2020] โ Raban Iten, Tony Metger, Henrik Wilming, Lรญdia Del Rio, and Renato Renner.Discovering physical concepts with neural networks.Phys. Rev. Lett., 124(1):010508, 2020. Cranmer et al. [2020] โ Miles D. Cranmer, Alvaro Sanchez-Gonzalez, Peter W. Battaglia, Rui Xu, Kyle Cranmer, David N. Spergel, and Shirley Ho.Discovering symbolic models from deep learning with inductive biases.In NeurIPS, 2020. Raissi et al. [2020] โ Maziar Raissi, Alireza Yazdani, and George Em Karniadakis.Hidden fluid mechanics: Learning velocity and pressure fields from flow visualizations.Science, 367(6481):1026โ1030, 2020. Raissi et al. [2019] โ Maziar Raissi, Paris Perdikaris, and George E. Karniadakis.Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations.J. Comput. Phys., 378:686โ707, 2019. Liu and Tegmark [2021] โ Ziming Liu and Max Tegmark.Machine learning conservation laws from trajectories.Phys. Rev. Lett., 126:180604, May 2021. Xue et al. [2021] โ Yexiang Xue, Md. Nasim, Maosen Zhang, Cuncai Fan, Xinghang Zhang, and Anter El-Azab.Physics knowledge discovery via neural differential equation embedding.In ECML/PKDD, pages 118โ134, 2021. Chen et al. [2018] โ Tian Qi Chen, Yulia Rubanova, Jesse Bettencourt, and David Duvenaud.Neural ordinary differential equations.In NeurIPS, pages 6572โ6583, 2018. Brence et al. [2021] โ Jure Brence, Ljupco Todorovski, and Saso Dzeroski.Probabilistic grammars for equation discovery.Knowl. Based Syst., 224:107077, 2021. Egan et al. [2024] โ Kevin Egan, Weizhen Li, and Rui Carvalho.Automatically discovering ordinary differential equations from data with sparse regression.Communications Physics, 7(1):20, 2024. Heinonen et al. [2018] โ Markus Heinonen, รagatay Yildiz, Henrik Mannerstrรถm, Jukka Intosalmi, and Harri Lรคhdesmรคki.Learning unknown ODE models with gaussian processes.In ICML, volume 80 of Proceedings of Machine Learning Research, pages 1964โ1973. PMLR, 2018. Li et al. [2021] โ Zongyi Li, Nikola Borislavov Kovachki, Kamyar Azizzadenesheli, Burigede Liu, Kaushik Bhattacharya, Andrew M. Stuart, and Anima Anandkumar.Fourier neural operator for parametric partial differential equations.In ICLR. OpenReview.net, 2021. Wagenmaker and Jamieson [2020] โ Andrew Wagenmaker and Kevin G. Jamieson.Active learning for identification of linear dynamical systems.In COLT, volume 125 of Proceedings of Machine Learning Research, pages 3487โ3582. PMLR, 2020. Mania et al. [2022] โ Horia Mania, Michael I. Jordan, and Benjamin Recht.Active learning for nonlinear system identification with guarantees.J. Mach. Learn. Res., 23:32:1โ32:30, 2022. Sener and Savarese [2018] โ Ozan Sener and Silvio Savarese.Active learning for convolutional neural networks: A core-set approach.In ICLR. OpenReview.net, 2018. Ash et al. [2020] โ Jordan T. Ash, Chicheng Zhang, Akshay Krishnamurthy, John Langford, and Alekh Agarwal.Deep batch active learning by diverse, uncertain gradient lower bounds.In ICLR. OpenReview.net, 2020. Haut et al. [2022] โ Nathan Haut, Wolfgang Banzhaf, and Bill Punch.Active learning improves performance on symbolic regression tasks in stackgp.In GECCO Companion, pages 550โ553, 2022. Lรผders et al. [2022] โ Christoph Lรผders, Thomas Sturm, and Ovidiu Radulescu.Odebase: a repository of ode systems for systems biology.Bioinformatics Advances, 2(1):vbac027, 2022. Kamienny et al. [2022] โ Pierre-Alexandre Kamienny, Stรฉphane dโAscoli, Guillaume Lample, and Franรงois Charton.End-to-end symbolic regression with transformers.In NeurIPS, 2022. Coddington and Levinson [1955] โ A. Coddington and N. Levinson.Theory of Ordinary Differential Equations.International series in pure and applied mathematics. McGraw-Hill, 1955.ISBN 9780070992566. Buisson-Fenet et al. [2020] โ Mona Buisson-Fenet, Friedrich Solowjow, and Sebastian Trimpe.Actively learning gaussian process dynamics.In L4DC, volume 120 of Proceedings of Machine Learning Research, pages 5โ15. PMLR, 2020. Langley [1977] โ Pat Langley.BACON: A production system that discovers empirical laws.In IJCAI, page 344, 1977. Langley [1979] โ Pat Langley.Rediscovering physics with BACON.3.In IJCAI, pages 505โ507, 1979. Langley et al. [1981] โ Pat Langley, Gary L. Bradshaw, and Herbert A. Simon.BACON.5: the discovery of conservation laws.In IJCAI, pages 121โ126, 1981. Kamienny et al. [2023] โ Pierre-Alexandre Kamienny, Guillaume Lample, Sylvain Lamprier, and Marco Virgolin.Deep generative symbolic regression with monte-carlo-tree-search.In ICML, volume 202. PMLR, 2023. Becker et al. [2023] โ Sรถren Becker, Michal Klein, Alexander Neitz, Giambattista Parascandolo, and Niki Kilbertus.Predicting ordinary differential equations with transformers.In ICML, volume 202 of Proceedings of Machine Learning Research, pages 1978โ2002. PMLR, 2023. Nagelkerke et al. [1991] โ Nico JD Nagelkerke et al.A note on a general definition of the coefficient of determination.Biometrika, 78(3):691โ692, 1991. Appendix AAvailability of Apps, Baselines and Dataset
Please find our code repository at: {mdframed} https://github.com/jiangnanhugo/ASD-ODE/
1.
the implementation of our Apps method is in the folder โapps_ode_pytorch/โ.
2.
the list of datasets is listed in โdata_oracle/scibench/scibench/data/โ.
3.
the implementation of several baseline algorithms is collected in folder โbaslines/โ.
We provide a โREADME.mdโ document for executing the programs.
We summarize the supplementary material as follows: Section B offers a detailed explanation of the phase portrait for ODE and the concepts of active learning; Section C provides the extended explanation of the proposed Apps method; Section D details the experimental settings; Section D.4 collects the extra experimental result.
Appendix BExtended Preliminaries
Phase Plane. The phase plane is a visual display of solutions of differential equations. Given an ODE, its solutions are a family of functions, which can be graphically plotted in the phase plane. At point ( ๐ฅ 1 , ๐ฅ 2 ) , we draw a vector representing the derivatives of the point with respect to the time variable, that is ( ๐ โข ๐ฅ 1 / ๐ โข ๐ก , ๐ โข ๐ฅ 2 / ๐ โข ๐ก ) . With sufficient of these arrows in place the system behavior over the regions of the place can be visualized and the long-term behavior can be quantitatively determined, like the limit cycles and attractors. The obtained entire figure is known as the phase portrait. It is a geometric representation of all possible trajectories from the corresponding ODE. One can interpret the phase plane in terms of dynamics. The solution of ODE corresponds to a trajectory of a point moving on the phase plane with velocity.
Butterfly Effect. In the context of ordinary differential equations (ODEs), the butterfly effect implies the sensitive dependence on initial conditions in dynamical systems. It means that small differences in the initial state of a system can lead to vastly different trajectories over time.
Mathematically, let ๐ฑ 0 and ๐ฑ 0 โฒ be two initial conditions that are very close to each other, i.e., โ ๐ฑ 0 โ ๐ฑ 0 โฒ โ โค ๐ฟ . The butterfly effect refers to the situation where the distance between the corresponding trajectories, ๐ฑ โข ( ๐ก ) and ๐ฑ โฒ โข ( ๐ก ) , grows exponentially over time:
โ ๐ฑ โข ( ๐ก ) โ ๐ฑ โฒ โข ( ๐ก ) โ โ exp โก ( ๐ โข ๐ก ) โข โ ๐ฑ 0 โ ๐ฑ 0 โฒ โ ,
where ๐ quantifies the rate at which two nearby trajectories diverge. If ๐ is positive, small initial differences grow exponentially over time, implying that even a tiny perturbation (like a butterfly flapping its wings) can lead to dramatically different outcomes in a chaotic system.
In this research, where the task is to query informative data to rank a given list of ODEs, this phenomenon implies the drawn data (i.e., initial conditions) are less informative but its close neighbor is highly informative. To avoid this, we can consider evaluating more initial conditions, which demand huge space usage and slow time computation.
Different trajectories never intersect in the phase plane. Solutions of ODEs are uniquely defined by initial conditions except at some special points in the phase plane. Trajectories in the phase plane cannot cross (except at some special points) as this would be equivalent to non-uniqueness of solutions. The special points are fixed points or singular points where trajectories start or end.
The existence and uniqueness theorem has an important corollary: different trajectories never intersect. If two trajectories did intersect, then there would be two solutions starting from the same point (the crossing point), and this would violate the uniqueness part of the theorem. In other words, a trajectory cannot move in two directions at once. Because trajectories cannot intersect, phase portraits always have a well-groomed look to them.
Theorem (Existence and Uniqueness [44]).
Consider the initial value problem ๐ฑ ห
๐ โข ( ๐ฑ ) , ๐ฑ โข ( 0 )
๐ฑ 0 . Suppose that ๐ is continuous and that all its partial derivatives โ ๐ ๐ / โ ๐ฅ ๐ , ๐ , ๐
1 , โฆ , ๐ , are continuous for ๐ฑ in some open connected set ๐ท โ โ ๐ . Then for ๐ฑ 0 โ ๐ท , the initial value problem has a solution ๐ฑ โข ( ๐ก ) on some time interval ( ๐ , ๐ ) about ๐ก
0 , and the solution is unique.
Active Learning is a machine learning method in which the learning algorithm inspects unlabeled data and interactively chooses the most informative data points to learn. The goal of active learning algorithms is to learn an accurate predictor with as little total data as possible. There are two standard settings: pool-based and streaming-based settings. In the pool-based setting, the learner is provided with a pool of unlabeled data, from which the interactively selects the most informative points and asks for their label. In the streaming-based setting, the learner receives a sequence of unlabeled points and decides on whether to request the label of the current point. In this research, we only consider pool-based active learning algorithms.
According to the active learning [12, 45, 41, 19], the input is the initial condition ๐ฑ 0 โ โ ๐ and the output is the obtained trajectory ( ๐ฑ ๐ก 1 , โฆ , ๐ฑ ๐ก ๐ ) โ โ ๐ ร ๐ . In the formulation of the Query-by-Committee method, they define the uncertainty at an input point as the variance of the predictions of the committee members.
Appendix CExtended Explanation of Apps method Data-availability Assumption.
A crucial assumption behind the success of Apps is the availability of a Data Oracle ๐ช that returns a (noisy) observation of the trajectory with a specified initial condition and a sequence of discrete times. Such a data oracle represents conducting controlled experiments in the real world, which can be expensive. This differs from the current symbolic regression, where a dataset is obtained prior to learning.
Vocabulary Construction.
Given the set of math operators and variables ๐ ๐
{ + , โ , ร , รท , sin โก โฆ } โช { ๐ฅ 1 , โฆ , ๐ฅ ๐ , ๐๐๐๐๐ } , where ๐๐๐๐๐ indidates the coefficients in the expressions. Following the definition of grammar in Section 3. For example, given ๐ ๐
{ + , โ , ร , รท } โช { ๐ฅ 1 , ๐ฅ 2 , ๐๐๐๐๐ } , we construct the following grammar rules:
A -> (A + A) A -> (A - A) A -> A * A A -> A / A A -> x1 A -> x2 A -> const B -> (B + B) B -> (B - B) B -> B * B B -> B / B B -> x1 B -> x2 B -> const
The non-terminal symbol โ ๐ด โ denotes a subexpression in ๐ โข ๐ฅ 1 and similarly โ ๐ต โ denotes a subexpression in ๐ โข ๐ฅ 2 . Each of them will be a unique token of the input vocabulary of the decoder and will be mapped to a distinct vector in the first embedding layer of the decoder. The above rules also form the output vocabulary of the decoder, where the neural decoder predicts a categorical distribution over the list of grammar rules.
The discovery path is to build algorithms that mimic human scientific discovery, which has achieved tremendous success in early works [46, 47, 48]. Recent work [21, 22, 41, 23] also pointed out the importance of having a data oracle that can actively query data points, rather than learning from a fixed dataset.
Sequential Decision Making Formulation.
We consider an undiscounted MDP with a finite horizon. The state space ๐ฎ is the set of all possible sequences of rules with maximum steps. The action space ๐ is the set of grammar rules. The ๐ก -th step state ๐ ๐ก is the sequence of sampled rules before the current step ๐ก , i.e., ๐ ๐ก := ( ๐ 1 , โฆ , ๐ ๐ก ) . The action ๐ ๐ก is the sampled single rule, ๐ ๐ก := ๐ ๐ก + 1 .
The loss function of Apps is informed by the REINFORCE algorithm [25], which is based on the log-derivative property:
โ ๐ ๐ ๐ โข ( ๐ )
๐ ๐ โข ( ๐ ) โข โ ๐ log โก ๐ ๐ โข ( ๐ )
where ๐ ๐ โข ( ๐ ) โ ( 0 , 1 ) represents a probability distribution over input ๐ with parameters ๐ and notation โ ๐ is the partial derivative with respect to ๐ . In our formulation, let ๐ ๐ โข ( ๐ ) denote the probability of sampling a sequence of grammar rules ๐ and ๐๐๐ ๐๐๐ โข ( ๐ )
1 / ( 1 + ๐ฝ๐ผ๐๐ด โข ( ๐ ) ) . Here ๐ is the corresponding expression constructed from the rules ๐ following the procedure in Section 3. The probability ๐ ๐ โข ( ๐ ) is modeled by the decoder modules. The objective described in Equation 1 is translated to maximize the expected reward of the sampled sequences from the decoder:
arg โก max ๐ โก ๐ผ ๐ โผ ๐ ๐ โข ( ๐ ) โข [ ๐๐๐ ๐๐๐ โข ( ๐ ) ]
Based on the REINFORCE algorithm, the gradient of the objective can be expanded as:
โ
๐
๐ผ
๐
โผ
๐
๐
โข
(
๐
)
[
๐๐๐ ๐๐๐
โข
(
๐
)
]
โ ๐ โข โ ๐ โ ฮฃ ๐๐๐ ๐๐๐ โข ( ๐ ) โข ๐ ๐ โข ( ๐ )
โ ๐ โ ฮฃ ๐๐๐ ๐๐๐ โข ( ๐ ) โข โ ๐ ๐ ๐ โข ( ๐ )
โ ๐ โ ฮฃ ๐๐๐ ๐๐๐ โข ( ๐ ) โข ๐ ๐ โข ( ๐ ) โข โ ๐ ๐ ๐ โข ( ๐ ) ๐ ๐ โข ( ๐ )
โ ๐ โ ฮฃ ๐๐๐ ๐๐๐ โข ( ๐ ) โข ๐ ๐ โข ( ๐ ) โข โ ๐ log โก ๐ ๐ โข ( ๐ )
๐ผ ๐ โผ ๐ ๐ โข ( ๐ ) โข [ ๐๐๐ ๐๐๐ โข ( ๐ ) โข โ ๐ log โก ๐ ๐ โข ( ๐ ) ]
where ฮฃ represents all possible sequences of grammar rules sampled from the decoder. The above expectation can be estimated by computing the averaged over samples drawn from the distribution ๐ ๐ โข ( ๐ ) . We first sample several times from the decoder module and obtain ๐ sequences ( ๐ 1 , โฆ , ๐ ๐ ) , an unbiased estimation of the gradient of the objective is computed as: โ ๐ ๐ฝ โข ( ๐ ) โ 1 ๐ โข โ ๐
1 ๐ ๐๐๐ ๐๐๐ โข ( ๐ ๐ ) โข โ ๐ log โก ๐ ๐ โข ( ๐ ๐ ) . In practice, the above computation has a high variance. To reduce variance, it is common to subtract a baseline function ๐ from the reward. In this study, we choose the baseline function as the average of the reward of the current sampled batch expressions. Thus we have:
โ
๐
๐ฝ
โข
(
๐
)
โ
1
๐
โข
โ
๐
1 ๐ ( ๐๐๐ ๐๐๐ โข ( ๐ ๐ ) โ ๐ ) โข โ ๐ log โก ๐ ๐ โข ( ๐ ๐ ) ,
where ๐
โ ๐
1 ๐ ๐๐๐ ๐๐๐ โข ( ๐ ๐ ) . Based on the description of the execution pipeline of the proposed Apps, we summarize every step in Algorithm 1.
Algorithm 1 Active Discovery of Ordinary Differential Equations via Phase Portrait Sketching. 1:Defined expression grammar; neural sequential decoder; data oracle for the ground-truth ODE ๐ช ; maximum learning epoch # โข ๐๐๐๐๐๐ ; discrete time steps ๐
( ๐ก 1 , โฆ , ๐ก ๐ ) . 2:The best-predicted ODE. 3:initialize the set of best predicted ODEs ๐ฌ โ โ ; โท initialization 4:randomly draw data ๐ท from Oracle ๐ช . 5:for ๐ก โ 1 โข to โข # โข ๐๐๐๐๐๐ do 6: sample ๐ sequences { ๐ 1 , โฆ , ๐ ๐ } from the sequential decoder. โท in Figure 2(a) 7: construct ๐ ODEs { ๐ ๐ } ๐
1 ๐ for each sequence from defined grammar. โท in Figure 2(b) 8: fit coefficients in each expression with data ๐ ๐ โ ๐พ๐๐๐๐๐๐ฃ๐ โข ( ๐ ๐ , ๐ท ) . 9: find a region ๐ข of high uncertainty with phase portrait sketching. โท in Figure 2(c) 10: draws trajectory data from Oracle in the selected region ๐ท ๐ข โ ๐ช โข ( ๐ข , ๐ ) . 11: computes reward using data ๐ท ๐ข . 12: save all expressions into ๐ฌ ; 13: save new data ๐ท ๐ข into ๐ท . 14: applies policy gradient to the parameters of the decoder. 15:return the expression in ๐ฌ with best goodness-of-fit on data ๐ท . C.1Implementation of Apps
In the experiments, we use an embedding layer, a multi-head self-attention layer, and finally softmax layer as the decoder. The dimension of the input embedding layer and the hidden vector is configured as 256 . We use the Adam optimizer as the gradient descent algorithm with a learning rate of 0.009 . The learning epoch is configured as 50 . The maximum sequence of grammar rules is fixed to be 20 . The batch size of expressions sampled from the decoder is set as 100 .
When fitting the values of coefficients in each expression, we sample a batch of data with batch size 1024 from the data Oracle. The open constants in the expressions are fitted on the data using the BFGS optimizer2.
We use a multi-processor Python library pathos to fit multiple expressions in parallel using 20 CPU cores. This greatly reduced the total time of the coefficients fitting step. The rest of the implementation details are available in the code implementation.
In terms of numerical integration, we use the fourth-order RungeโKutta (RK45) method to compute the trajectory data of the specified ODEs. Other numerical integration algorithms in sicpy.integrate are implemented with adaptive time step size, which makes it very slow when fitting the coefficients in the candidate ODEs.
Every candidate ODE is represented by the Sympy Lambdify function. We implement the Runge-Kutta function instead of using the Scipy.integrate.solve_ivp function, where the latter has internally used the adaptive step size and is extremely slow for specific candidate ODEs. In our experiments, we find that the Scipy.integrate.solve_ivp API cannot return a trajectory when running together with the coefficient fitting steps for more than 12 hours.
Figure 4:Implemented 4th order Runge Kutter method.
In Figure 4, given one predicted ODE expr_strs, which is represented as an array of expressions of length ๐ , the following lines of code compute the predicted trajectory pred_trajectories for a batch of initial conditions x_init_conds. input_var_Xs the list of symbols of variables: [ ๐ฅ 1 , โฆ , ๐ฅ ๐ ] . We can then compare if the predicted trajectories are close to the ground-truth trajectories using the NMSE metric, given the same set of initial conditions.
Hyper-parameter Configuration
An expression containing placeholder symbol ๐ด or containing more than 20 open constants is not evaluated on the data, the fitness score of it is โ โ . In terms of the reward function in the policy gradient objective, we use ๐๐๐ ๐๐๐ โข ( ๐ )
1 1 + ๐ฝ๐ผ๐๐ด โข ( ๐ ) . The normalized mean-squared error metric is further defined in Equation 4. We set the relative size of the interval of the region to be 1 / 4 , meaning the length of every edge of the region is 1 / 4 of the original interval for each variable. We set the number of regions to be 10 . The region is randomly generated by first generating the leftmost point in the original variable interval and then applying each edge of the region with a given relative length.
The deep network part is implemented using the most recent version of Pytorch, the expression evaluation is based on the Sympy library, and the step for fitting open constants in expression with the dataset uses the Scipy library. The visualization of the phase portrait is from 3 and the part of the experimental visualization is borrowed from 4.
C.2Limitation and Broader Impact
Limitation The proposed Apps that generates phase portraits can suffer from resolution issues. Fine details of the dynamics might be missed if the region width is too large or the number of sampled points in each region is too small. It is also unclear if the proposed phase portrait sketching idea is applicable to partial differential equations.
Broad Impact The proposed Apps can be useful for scientists to actively discover governing laws from data. It will accelerate the discovery process compared to passive learning algorithms.
Appendix DExperiment Settings D.1Baselines
For the baselines of the ODE discovery task, we consider
โข
SINDy [5]5 is a popular method using a sparse regression algorithm to find the differential equations.
โข
ProGED [33]6 uses probabilistic context-free grammar to search for differential equations. ProGED first samples a list of candidate expressions from the defined probabilistic context-free grammar for symbolic expressions. Then ProGED fits the open constants in each expression using the given training dataset. The equation with the best fitness scores is returned.
โข
ODEFormer [10]7 is the most recent framework that uses the transformer for the discovery of ordinary differential equations. We use the provided pre-trained model to predict the governing expression with the dataset. We execute the model 10 times and pick the expression with the smallest NMSE error. The dataset size is 500 , which is the largest dataset configuration for the ODEFormer.
In principle, the ODE discovery task can be formulated as a symbolic regression task where the input is ๐ฑ ๐ก and the output is directly ๐ฑ ห ๐ก . Given the trajectory data ( ๐ฑ 0 , ๐ฑ โข ( ๐ก 1 ) , โฆ , ๐ฑ โข ( ๐ก ๐ ) ) , its output label is approximated by computing:
๐ฑ ห โข ( ๐ก ๐ ) โ ( ๐ฑ โข ( ๐ก ๐ + 1 ) โ ๐ฑ โข ( ๐ก ๐ ) ) ( ๐ก ๐ + 1 โ ๐ก ๐ )
In literature, this approach is called symbolic regression with gradient matching. We consider two representative baselines in the symbolic regression task:
โข
Symbolic Physics Learner (SPL) is a heuristic search algorithm based on Monte Carlo Tree Search for finding optimal sequences of production rules using context-free grammars [49, 7]8. It employs Monte Carlo simulations to explore the search space of all the production rules and determine the value of each node in the search tree. SPL consists of four steps in each iteration: 1) Selection. Starting at a root node, recursively select the optimal child (i.e., one of the production rules) until reaching an expandable node or a leaf node. 2) Expansion. If the expandable node is not the terminal, create one or more of its child nodes to expand the search tree. 3) Simulation. Run a simulation from the new node until achieving the result. 4) Backpropagation. Update the node sequence from the new node to the root node with the simulated result. To balance the selection of optimal child node(s) by exploiting known rewards (exploitation) or expanding a new node to explore potential rewards exploration, the upper confidence bound (UCB) is often used.
โข
End to End Transformer for symbolic regression (E2ETransformer) [43]9. They propose to use a deep transformer to pre-train on a large set of randomly generated expressions. We load the shared pre-trained model. We provide the given dataset and the E2ETransformer infers 10 best expressions. We choose to report the expression with the best NMSE scores.
Note that the open link in NSODE [50] has no code implementation by far. Thus, the NSODE approach is not included for comparison. Also, a recent method [19] proposes a new learning framework for actively drawing data. However, their approaches are designed for searching algebraic equations, and the integration into differential equations is unclear. So their approach is also not considered for comparison in this work.
Apps ProGED ODEFormer SPL
goodness-of-fit function NegMSE NegMSE NegRMSE NegRMSE Training setting evolve for 1 sec with time step 0.001 sec and 100 random initial conditions Testing setting evolve for 10 sec with time step 0.001 sec and 100 random initial conditions #CPU for training 20 Table 4:Major hyper-parameters settings for all the algorithms considered in the experiment. D.2Evaluation Metrics
The goodness-of-fit indicates how well the learning algorithms perform in discovering unknown ODEs. The testing set contains new trajectories with new initial conditions, generated from the ground-truth ODE. we measure the goodness-of-fit of a predicted expression ๐
[ ๐ 1 , โฆ , ๐ ๐ ] , by evaluating the mean-squared-error (MSE) and normalized-mean-squared-error (NMSE):
NMSE
= 1 ๐ 2 โข 1 ๐ โข โ ๐
1 ๐ ( ๐ฑ โข ( ๐ก ๐ ) โ ๐ฑ ^ โข ( ๐ก ๐ ) ) 2 ,
(4)
The empirical variance ๐
1 ๐ โข โ ๐
1 ๐ ( ๐ฑ ๐ โ 1 ๐ โข โ ๐
1 ๐ ๐ฑ ๐ ) 2 . We use the NMSE as the main criterion for comparison in the experiments and present the results on the remaining metrics in the case studies. The main reason is that the NMSE is less impacted by the output range. The output ranges of expression are dramatically different from each other, making it difficult to present results uniformly if we use other metrics.
Prior work [15] further proposed a coefficient of determination ๐ 2 -based score over a group of expressions in the dataset, as a statistical measure of whether the best-predicted expression is almost close to the ground-truth expression. An ๐ 2 of 1 indicates that the regression predictions perfectly fit the data [51]. The ๐ 2 score is computed as follows:
๐ 2 โข ( ๐ ๐ )
1 โ NMSE โข ( ๐ ๐ )
(5) D.3Computational Resource
All the methods are running with Python 3.10 and on the same set of hardware where the CPU is Milan CPUs @ 2.45GHz, the RAM is set as 8GB, and the maximum running time is set as 24 hours. The extra necessary configuration is collected in the anonymous code repository. The hyper-parameter configurations of baselines are listed in Table 4.
D.4Extended Experimental Results
The given set of best-predicted ODEs for Table 3 is shown in Figure 5.
๐ 1
(
0.088
โข
๐ฅ
1
,
0.146
โข
cos
โก
(
๐ฅ
1
)
)
๐
2
(
โ
0.0107
โข
sin
โก
(
๐ฅ
1
)
,
0.712
)
๐
3
(
2
/
cos
โก
(
๐ฅ
1
)
+
0.1759
โข
sin
โก
(
๐ฅ
1
)
,
0.820
)
๐
4
(
0.645
sin
(
๐ฅ
0
)
โ
0.893
cos
(
๐ฅ
0
)
,
0.2891
๐ฅ
1
(
๐ฅ
1
โ
0.153
โข
cos
โก
(
๐ฅ
0
)
)
+
0.213
sin
โก
(
๐ฅ
0
)
(
๐ฅ
1
โ
0.153
โข
cos
โก
(
๐ฅ
0
)
)
โ
0.0443
cos
โก
(
๐ฅ
0
)
(
๐ฅ
1
โ
0.15
โข
cos
โก
(
๐ฅ
0
)
)
)
๐
5
(
0.95
โข
sin
โก
(
๐ฅ
0
)
โ
0.29
,
0.14153
โข
๐ฅ
1
โ
0.691
โข
sin
โก
(
๐ฅ
1
)
cos
โก
(
๐ฅ
0
)
โ
0.0673
โข
cos
โก
(
๐ฅ
0
)
+
0.8825
โข
cos
โก
(
๐ฅ
1
)
cos
โก
(
๐ฅ
0
)
)
๐
6
(
1.63
โข
sin
โก
(
๐ฅ
0
)
,
๐ฅ
0
)
๐
7
(
1.15
โข
sin
โก
(
๐ฅ
0
)
,
โ
70.56
โข
cos
โก
(
๐ฅ
1
)
)
๐
8
(
0.71
,
โ
70.56
โข
cos
โก
(
๐ฅ
1
)
)
๐
9
(
0.99
โข
cos
โก
(
๐ฅ
0
)
,
โ
1.0
โข
๐ฅ
0
โ
39.53
โข
sin
โก
(
๐ฅ
1
)
โ
152.57
โข
cos
โก
(
๐ฅ
1
)
)
๐
10
(
โ
0.871
โข
sin
โก
(
๐ฅ
1
)
,
โ
3.826
โข
sin
โก
(
๐ฅ
0
)
โ
0.212
โข
sin
โก
(
๐ฅ
1
)
+
4.311
)
๐
11
(
0.088
โข
๐ฅ
1
,
0.146
โข
cos
โก
(
๐ฅ
1
)
)
๐
12
(
โ
0.010
โข
sin
โก
(
๐ฅ
1
)
,
0.712
)
๐
13
(
1.477
โข
๐
โ
5
โข
๐ฅ
1
โข
2
/
cos
โก
(
๐ฅ
1
)
+
0.1759
โข
sin
โก
(
๐ฅ
1
)
,
0.8205
)
๐
14
(
0.645
โข
sin
โก
(
๐ฅ
0
)
โ
0.893
โข
cos
โก
(
๐ฅ
0
)
,
0.289
โข
๐ฅ
1
(
๐ฅ
1
โ
0.15
โข
cos
โก
(
๐ฅ
0
)
)
+
0.21
โข
sin
โก
(
๐ฅ
0
)
(
๐ฅ
1
โ
0.153
โข
cos
โก
(
๐ฅ
0
)
)
โ
0.044
โข
cos
โก
(
๐ฅ
0
)
(
๐ฅ
1
โ
0.1532
โข
cos
โก
(
๐ฅ
0
)
)
)
๐
15
(
0.953
โข
sin
โก
(
๐ฅ
0
)
โ
0.2922
,
0.1415
โข
๐ฅ
1
โ
0.691
โข
sin
โก
(
๐ฅ
1
)
/
cos
โก
(
๐ฅ
0
)
โ
0.0673
โข
cos
โก
(
๐ฅ
0
)
+
0.882
โข
cos
โก
(
๐ฅ
1
)
cos
โก
(
๐ฅ
0
)
)
๐
16
(
1.631
โข
sin
โก
(
๐ฅ
0
)
,
๐ฅ
0
)
๐
17
(
1.152
โข
sin
โก
(
๐ฅ
0
)
,
โ
70.56
โข
cos
โก
(
๐ฅ
1
)
)
๐
18
(
0.710
,
โ
70.56
โข
cos
โก
(
๐ฅ
1
)
)
๐
19
(
0.99
cos
(
๐ฅ
0
)
,
โ
1.0
๐ฅ
0
โ
39.53
sin
(
๐ฅ
1
)
โ
152.57
cos
(
๐ฅ
1
)
)
๐
20
( โ 0.871 โข sin โก ( ๐ฅ 1 ) , โ 3.826 โข sin โก ( ๐ฅ 0 ) โ 0.211 โข sin โก ( ๐ฅ 1 ) + 4.3117 )
Figure 5:The given set of best-predicted ODEs for Table 3.
The Kendall tau distance is computed with two ranked lists, one of them is evaluated on full data and another is evaluated on the chosen region with high uncertainty. We use library10 to compute the ranking score.
D.5Dataset
We present the visualized phase portraits and the explicit forms of the ODEs considered in this research in the following figures and tables.
Figure 6:Selected phase portrait of Strogatz dataset with variables ๐
2 . Figure 7:Selected phase portrait of Strogatz dataset with variables ๐
3
.
ID Equation
1 RC-circuit (charging capacitor)
๐ฅ
ห
0
( 0.7 โ ๐ฅ 0 / 1.2 ) / 2.31
2 Population growth (naive)
๐ฅ
ห
0
0.23 โข ๐ฅ 0
3 Population growth with carrying capacity
๐ฅ
ห
0
0.79 โข ๐ฅ 0 โข ( 1 โ ๐ฅ 0 / 74.3 )
4 RC-circuit with non-linear resistor (charging capacitor)
๐ฅ
ห
0
1 / ( 1 + exp โก ( 0.5 โ ๐ฅ 0 / 0.96 ) ) โ 0.5
5 Velocity of a falling object with air resistance
๐ฅ
ห
0
9.81 โ 0.0021175 โข ๐ฅ 0 2
6 Autocatalysis with one fixed abundant chemical
๐ฅ
ห
0
2.1 โข ๐ฅ 0 โ 0.5 โข ๐ฅ 0 2
7 Gompertz law for tumor growth
๐ฅ
ห
0
0.032 โข ๐ฅ 0 โข log โก ( 2.29 โข ๐ฅ 0 )
8 Logistic equation with Allee effect
๐ฅ
ห
0
0.14 โข ๐ฅ 0 โข ( 1 โ ๐ฅ 0 / 130.0 ) โข ( ๐ฅ 0 / 4.4 โ 1 )
9 Language death model for two languages
๐ฅ
ห
0
( 1 โ ๐ฅ 0 ) โข 0.32 โ ๐ฅ 0 โข 0.28
10 Refined language death model for two languages
๐ฅ
ห
0
( 1 โ ๐ฅ 0 ) โข 0.2 โข ๐ฅ 0 1.2 โ ๐ฅ 0 โข ( 1 โ 0.2 ) โข ( 1 โ ๐ฅ 0 ) 1.2
11 Naive critical slowing down (statistical mechanics)
๐ฅ
ห
0
โ ๐ฅ 0 3
12 Photons in a laser (simple)
๐ฅ
ห
0
1.8 โข ๐ฅ 0 โ 0.1107 โข ๐ฅ 0 2
13 Overdamped bead on a rotating hoop
๐ฅ
ห
0
0.0981 โข sin โก ( ๐ฅ 0 ) โข ( 9.7 โข cos โก ( ๐ฅ 0 ) โ 1 )
14 Budworm outbreak model with predation
๐ฅ
ห
0
0.78 โข ๐ฅ 0 โข ( 1 โ ๐ฅ 0 / 81.0 ) โ 0.9 โข ๐ฅ 0 2 / ( 21.2 2 + ๐ฅ 0 2 )
15 Budworm outbreak with predation (dimensionless)
๐ฅ
ห
0
0.4 โข ๐ฅ 0 โข ( 1 โ ๐ฅ 0 / 95.0 ) โ ๐ฅ 0 2 / ( 1 + ๐ฅ 0 2 )
16 Landau equation (typical time scale tau = 1)
๐ฅ
ห
0
0.1 ๐ฅ 0 โ โ 0.04 ๐ฅ 0 3 โ 0.001 ๐ฅ 0 5
17 Logistic equation with harvesting/fishing
๐ฅ
ห
0
0.4 โข ๐ฅ 0 โข ( 1 โ ๐ฅ 0 / 100.0 ) โ 0.3
18 Improved logistic equation with harvesting/fishing
๐ฅ
ห
0
0.4 โข ๐ฅ 0 โข ( 1 โ ๐ฅ 0 / 100.0 ) โ 0.24 โข ๐ฅ 0 / ( 50.0 + ๐ฅ 0 )
19 Improved logistic equation with harvesting/fishing (dimensionless)
๐ฅ
ห
0
๐ฅ 0 โข ( 1 โ ๐ฅ 0 ) โ 0.08 โข ๐ฅ 0 / ( 0.8 + ๐ฅ 0 )
20 Autocatalytic gene switching (dimensionless)
๐ฅ
ห
0
0.1 โ 0.55 โข ๐ฅ 0 + ๐ฅ 0 2 / ( 1 + ๐ฅ 0 2 )
21 Dimensionally reduced SIR infection model for dead people (dimensionless)
๐ฅ
ห
0
1.2 โ 0.2 โข ๐ฅ 0 โ exp โก ( โ ๐ฅ 0 ) Table 5:Selected Strogatz dataset with variable ๐
1
.
ID Explicit Equations
1 Harmonic oscillator without damping
๐ฅ
ห
0
๐ฅ
1
๐ฅ
ห
1
โ 2.1 โข ๐ฅ 0
2 Harmonic oscillator with damping
๐ฅ
ห
0
๐ฅ
1
๐ฅ
ห
1
โ 4.5 โข ๐ฅ 0 โ 0.43 โข ๐ฅ 1
3 Lotka-Volterra competition model (Strogatz version with sheeps and rabbits)
๐ฅ
ห
0
๐ฅ
0
โข
(
3.0
โ
๐ฅ
0
โ
2.0
โข
๐ฅ
1
)
๐ฅ
ห
1
๐ฅ 1 โข ( 2.0 โ ๐ฅ 0 โ ๐ฅ 1 )
4 Lotka-Volterra simple (as on Wikipedia)
๐ฅ
ห
0
๐ฅ
0
โข
(
1.84
โ
1.45
โข
๐ฅ
1
)
๐ฅ
ห
1
โ ๐ฅ 1 โข ( 3.0 โ 1.62 โข ๐ฅ 0 )
5 Pendulum without friction
๐ฅ
ห
0
๐ฅ
1
๐ฅ
ห
1
โ 0.9 โข sin โก ( ๐ฅ 0 )
6 Dipole fixed point
๐ฅ
ห
0
0.65
โข
๐ฅ
0
โข
๐ฅ
1
๐ฅ
ห
1
๐ฅ 1 2 โ ๐ฅ 0 2
7 RNA molecules catalyzing each others replication
๐ฅ
ห
0
๐ฅ
0
โข
(
๐ฅ
1
โ
1.61
โข
๐ฅ
0
โข
๐ฅ
1
)
๐ฅ
ห
1
๐ฅ 1 โข ( ๐ฅ 0 โ 1.61 โข ๐ฅ 0 โข ๐ฅ 1 )
8 SIR infection model only for healthy and sick
๐ฅ
ห
0
โ
0.4
โข
๐ฅ
0
โข
๐ฅ
1
๐ฅ
ห
1
0.4 โข ๐ฅ 0 โข ๐ฅ 1 โ 0.314 โข ๐ฅ 1
9 Damped double well oscillator
๐ฅ
ห
0
๐ฅ
1
๐ฅ
ห
1
โ 0.18 โข ๐ฅ 1 + ๐ฅ 0 โ ๐ฅ 0 3
10 Glider (dimensionless)
๐ฅ
ห
0
โ
sin
โก
(
๐ฅ
1
)
โ
0.08
โข
๐ฅ
0
2
๐ฅ
ห
1
๐ฅ 0 โ cos โก ( ๐ฅ 1 ) / ๐ฅ 0
11 Frictionless bead on a rotating hoop (dimensionless)
๐ฅ
ห
0
๐ฅ
1
๐ฅ
ห
1
sin โก ( ๐ฅ 0 ) โข ( cos โก ( ๐ฅ 0 ) โ 0.93 )
12 Rotational dynamics of an object in a shear flow
๐ฅ
ห
0
cot
โก
(
๐ฅ
1
)
โข
cos
โก
(
๐ฅ
0
)
๐ฅ
ห
1
sin ( ๐ฅ 0 ) ( cos ( ๐ฅ 1 ) 2 + 4.2 sin ( ๐ฅ 1 ) 2 )
13 Pendulum with non-linear damping, no driving (dimensionless)
๐ฅ
ห
0
๐ฅ
1
๐ฅ
ห
1
โ sin โก ( ๐ฅ 0 ) โ ๐ฅ 1 โ 0.07 โข cos โก ( ๐ฅ 0 ) โข ๐ฅ 1
14 Van der Pol oscillator (standard form)
๐ฅ
ห
0
๐ฅ
1
๐ฅ
ห
1
โ ๐ฅ 0 โ 0.43 โข ( ๐ฅ 0 2 โ 1 ) โข ๐ฅ 1
15 Van der Pol oscillator (simplified form from Strogatz)
๐ฅ
ห
0
3.37
โข
(
๐ฅ
1
โ
๐ฅ
0
3
/
3
+
๐ฅ
0
)
๐ฅ
ห
1
โ ๐ฅ 0 / 3.37
16 Glycolytic oscillator, e.g., ADP and F6P in yeast (dimensionless)
๐ฅ
ห
0
โ
๐ฅ
0
+
2.4
โข
๐ฅ
1
+
๐ฅ
0
2
โข
๐ฅ
1
๐ฅ
ห
1
0.07 โ 2.4 โข ๐ฅ 0 โ ๐ฅ 0 2 โข ๐ฅ 1
17 Duffing equation (weakly non-linear oscillation)
๐ฅ
ห
0
๐ฅ
1
๐ฅ
ห
1
โ ๐ฅ 0 + 0.886 โข ๐ฅ 1 โข ( 1 โ ๐ฅ 0 2 )
18 Cell cycle model by Tyson for interaction between protein cdc2and cyclin (dimensionless)
๐ฅ
ห
0
15.3
โข
(
๐ฅ
1
โ
๐ฅ
0
)
โข
(
0.001
+
๐ฅ
0
2
)
โ
๐ฅ
0
๐ฅ
ห
1
0.3 โ ๐ฅ 0
19 Reduced model for chlorine dioxide-iodine-malonic acid rection (dimensionless)
๐ฅ
ห
0
8.9
โ
๐ฅ
0
โ
4.0
โข
๐ฅ
0
โข
๐ฅ
1
/
(
1
+
๐ฅ
0
2
)
๐ฅ
ห
1
1.4 โข ๐ฅ 0 โข ( 1 โ ๐ฅ 1 / ( 1 + ๐ฅ 0 2 ) )
20 Driven pendulum with linear damping / Josephson junction (dimensionless)
๐ฅ
ห
0
๐ฅ
1
๐ฅ
ห
1
1.67 โ sin โก ( ๐ฅ 0 ) โ 0.64 โข ๐ฅ 1
21 Driven pendulum with quadratic damping (dimensionless)
๐ฅ
ห
0
๐ฅ
1
๐ฅ
ห
1
1.67 โ sin โก ( ๐ฅ 0 ) โ 0.64 โข ๐ฅ 1 1 Table 6:Selected Strongatz dataset with variables ๐
2
.
ID Explicit Equations
1 Maxwell-Bloch equations (laser dynamics)
๐ฅ
ห
0
0.1 โข ( ๐ฅ 1 โ ๐ฅ 0 )
๐ฅ ห 1
0.21 โข ( ๐ฅ 0 โข ๐ฅ 2 โ ๐ฅ 1 )
๐ฅ ห 2
0.34 โข ( 3.1 + 1 โ ๐ฅ 2 โ 3.1 โข ๐ฅ 0 โข ๐ฅ 1 )
2 Model for apoptosis (cell death)
๐ฅ
ห
0
0.1 โ 0.4 โข ๐ฅ 1 โข ๐ฅ 0 / ( 0.1 + ๐ฅ 0 ) โ 0.05 โข ๐ฅ 0
๐ฅ ห 1
0.6 โข ๐ฅ 2 โข ( 0.1 + ๐ฅ 1 ) โ 0.2 โข ๐ฅ 1 / ( 0.1 + ๐ฅ 1 ) โ 7.95 โข ๐ฅ 0 โข ๐ฅ 1 / ( 2.0 + ๐ฅ 1 )
๐ฅ ห 2
โ 0.6 โข ๐ฅ 2 โข ( 0.1 + ๐ฅ 1 ) + 0.2 โข ๐ฅ 1 / ( 0.1 + ๐ฅ 1 ) + 7.95 โข ๐ฅ 0 โข ๐ฅ 1 / ( 2.0 + ๐ฅ 1 )
3 Lorenz equations in well-behaved periodic regime
๐ฅ
ห
0
5.1 โข ( ๐ฅ 1 โ ๐ฅ 0 )
๐ฅ ห 1
12.0 โข ๐ฅ 0 โ ๐ฅ 1 โ ๐ฅ 0 โข ๐ฅ 2
๐ฅ ห 2
๐ฅ 0 โข ๐ฅ 1 โ 1.67 โข ๐ฅ 2
4 Lorenz equations in complex periodic regime
๐ฅ
ห
0
10.0 โข ( ๐ฅ 1 โ ๐ฅ 0 )
๐ฅ ห 1
99.96 โข ๐ฅ 0 โ ๐ฅ 1 โ ๐ฅ 0 โข ๐ฅ 2
๐ฅ ห 2
๐ฅ 0 โข ๐ฅ 1 โ 2.6666666666666665 โข ๐ฅ 2
5 Lorenz equations standard parameters (chaotic)
๐ฅ
ห
0
10.0 โข ( ๐ฅ 1 โ ๐ฅ 0 )
๐ฅ ห 1
28.0 โข ๐ฅ 0 โ ๐ฅ 1 โ ๐ฅ 0 โข ๐ฅ 2
๐ฅ ห 2
๐ฅ 0 โข ๐ฅ 1 โ 2.6666666666666665 โข ๐ฅ 2
6 Rรถssler attractor (stable fixed point)
๐ฅ
ห
0
5.0 โข ( โ ๐ฅ 1 โ ๐ฅ 2 )
๐ฅ ห 1
5.0 โข ( ๐ฅ 0 โ 0.2 โข ๐ฅ 1 )
๐ฅ ห 2
5.0 โข ( 0.2 + ๐ฅ 2 โข ( ๐ฅ 0 โ 5.7 ) )
7 Rรถssler attractor (periodic)
๐ฅ
ห
0
5.0 โข ( โ ๐ฅ 1 โ ๐ฅ 2 )
๐ฅ ห 1
5.0 โข ( ๐ฅ 0 + 0.1 โข ๐ฅ 1 )
๐ฅ ห 2
5.0 โข ( 0.2 + ๐ฅ 2 โข ( ๐ฅ 0 โ 5.7 ) )
8 Rรถssler attractor (chaotic)
๐ฅ
ห
0
5.0 โข ( โ ๐ฅ 1 โ ๐ฅ 2 )
๐ฅ ห 1
5.0 โข ( ๐ฅ 0 + 0.2 โข ๐ฅ 1 )
๐ฅ ห 2
5.0 โข ( 0.2 + ๐ฅ 2 โข ( ๐ฅ 0 โ 5.7 ) )
9 Aizawa attractor (chaotic)
๐ฅ
ห
0
๐ฅ 0 โข ( ๐ฅ 2 โ 0.7 ) โ 3.5 โข ๐ฅ 1
๐ฅ ห 1
3.5 โข ๐ฅ 0 + ๐ฅ 1 โข ( ๐ฅ 2 โ 0.7 )
๐ฅ ห 2
0.65 + 0.95 โข ๐ฅ 2 โ ๐ฅ 2 3 / 3 . โ ( ๐ฅ 0 2 + ๐ฅ 1 2 ) โข ( 1 + 0.25 โข ๐ฅ 2 ) + 0.1 โข ๐ฅ 2 โข ๐ฅ 0 3
10 Chen-Lee attractor
๐ฅ
ห
0
5 โข ๐ฅ 0 โ ๐ฅ 1 โข ๐ฅ 2
๐ฅ ห 1
โ 10.0 โข ๐ฅ 1 + ๐ฅ 0 โข ๐ฅ 2
๐ฅ ห 2
โ 3.8 โข ๐ฅ 2 + ๐ฅ 0 โข ๐ฅ 1 / 3.0 Table 7:The Strogatz dataset with variables ๐
3
.
ID Equations
1 Norel1990 - MPF and Cyclin Oscillations
๐ฅ
ห
0
1.0 โข ๐ฅ 0 2 โข ๐ฅ 1 โ 10.0 โข ๐ฅ 0 / ( ๐ฅ 0 + 1.0 ) + 3.466 โข ๐ฅ 1
๐ฅ ห 1
1.2 โ 1.0 โข ๐ฅ 0
2 Chrobak2011 - A mathematical model of induced cancer-adaptive immune system competition
๐ฅ
ห
0
โ 0.03125 โข ๐ฅ 0 2 โ 0.125 โข ๐ฅ 0 โข ๐ฅ 1 + 0.0625 โข ๐ฅ 0
๐ฅ ห 1
โ 0.08594 โข ๐ฅ 0 โข ๐ฅ 1 โ 0.03125 โข ๐ฅ 1 2 + 0.03125 โข ๐ฅ 1
3 FitzHugh1961-NerveMembrane
๐ฅ
ห
0
โ 1.0 โข ๐ฅ 0 3 + 3.0 โข ๐ฅ 0 + 3.0 โข ๐ฅ 1 โ 1.2
๐ฅ ห 1
โ 0.3333 โข ๐ฅ 0 โ 0.2667 โข ๐ฅ 1 + 0.2333
4 Clarke2000 - One-hit model of cell death in neuronal degenerations
๐ฅ
ห
0
โ 0.278 โข ๐ฅ 0
๐ฅ ห 1
โ 0.223 โข ๐ฅ 1
5 Wodarz2018/1 - simple model
๐ฅ
ห
0
0.004 โข ๐ฅ 0 + 0.004 โข ๐ฅ 1 / ( 0.01 โข ๐ฅ 0 1 โข .0 + 1.0 )
๐ฅ ห 1
0.006 โข ๐ฅ 0 โ 0.003 โข ๐ฅ 1 โ 0.004 โข ๐ฅ 1 / ( 0.01 โข ๐ฅ 0 1 โข .0 + 1.0 )
6 Ehrenstein2000 - Positive-Feedback model for the loss of acetylcholine in Alzheimerโs disease
๐ฅ
ห
0
โ 0.007 โข ๐ฅ 0 โข ๐ฅ 1
๐ฅ ห 1
โ 0.004 โข ๐ฅ 0 โ 0.01 โข ๐ฅ 1 + 0.33
7 Cao2013 - Application of ABSIS method in the bistable Schlagl model
๐ฅ
ห
0
0.12 โข ๐ฅ 0 2 โ 3.071 โข ๐ฅ 0 + 12.5 โ 0.00192 / ๐ฅ 0
๐ฅ ห 1
โ 0.12 โข ๐ฅ 0 2 + 3.071 โข ๐ฅ 0 โ 12.5 + 0.00192 / ๐ฅ 0
8 Chaudhury2020 - Lotka-Volterra mathematical model of CAR-T cell and tumour kinetics
๐ฅ
ห
0
0.002 โข ๐ฅ 0 โข ๐ฅ 1 โ 0.16 โข ๐ฅ 0
๐ฅ ห 1
0.15 โข ๐ฅ 1
9 Baker2013 - Cytokine Mediated Inflammation in Rheumatoid Arthritis
๐ฅ
ห
0
โ ๐ฅ 0 + 3.5 โข ๐ฅ 1 2 / ( ๐ฅ 1 2 + 0.25 )
๐ฅ ห 1
1.0 โข ๐ฅ 1 2 / ( ๐ฅ 0 2 โข ๐ฅ 1 2 + ๐ฅ 0 2 + ๐ฅ 1 2 + 1.0 ) โ 1.25 โข ๐ฅ 1 + 0.025 / ( ๐ฅ 0 2 + 1.0 )
10 Somogyi1990-CaOscillations
๐ฅ
ห
0
โ 5.0 โข ๐ฅ 0 โข ๐ฅ 1 4 โข .0 / ( ๐ฅ 1 4 โข .0 + 81.0 ) โ 0.01 โข ๐ฅ 0 + 2.0 โข ๐ฅ 1
๐ฅ ห 1
5.0 โข ๐ฅ 0 โข ๐ฅ 1 4 โข .0 / ( ๐ฅ 1 4 โข .0 + 81.0 ) + 0.01 โข ๐ฅ 0 โ 3.0 โข ๐ฅ 1 + 1.0
11 Cucuianu2010 - A hypothetical-mathematical model of acute myeloid leukaemia pathogenesis
๐ฅ
ห
0
โ 0.1 โข ๐ฅ 0 + 0.3 โข ๐ฅ 0 / ( 0.5 โข ๐ฅ 0 + 0.5 โข ๐ฅ 1 + 1.0 )
๐ฅ ห 1
โ 0.1 โข ๐ฅ 1 + 0.3 โข ๐ฅ 1 / ( 0.5 โข ๐ฅ 0 + 0.5 โข ๐ฅ 1 + 1.0 )
12 Wang2016/3 - oncolytic efficacy of M1 virus-SN model
๐ฅ
ห
0
โ 0.2 โข ๐ฅ 0 โข ๐ฅ 1 โ 0.02 โข ๐ฅ 0 + 0.02
๐ฅ ห 1
0.16 โข ๐ฅ 0 โข ๐ฅ 1 โ 0.03 โข ๐ฅ 1
13 Chen2011/1 - bone marrow invasion absolute model
๐ฅ
ห
0
โ 0.2 โข ๐ฅ 0 2 + 0.1 โข ๐ฅ 0
๐ฅ ห 1
โ 1.0 โข ๐ฅ 0 โข ๐ฅ 1 โ 0.8 โข ๐ฅ 1 2 + 0.7 โข ๐ฅ 1
14 Cao2013 - Application of ABSIS method in the reversible isomerization model
๐ฅ
ห
0
โ 0.12 โข ๐ฅ 0 + 1.0 โข ๐ฅ 1
๐ฅ ห 1
0.12 โข ๐ฅ 0 โ 1.0 โข ๐ฅ 1 Table 8:Selected ODEBase dataset with variables ๐
2
.
1 Turner2015-Human/Mosquito ELP Model
๐ฅ
ห
0
600.0 โ 0.411 โข ๐ฅ 0
๐ฅ ห 1
0.361 โข ๐ฅ 0 โ 0.184 โข ๐ฅ 1
๐ฅ ห 2
0.134 โข ๐ฅ 1 โ 0.345 โข ๐ฅ 2
2 Al-Tuwairqi2020 - Dynamics of cancer virotherapy - Phase I treatment
๐ฅ
ห
0
โ 1.0 โข ๐ฅ 0 โข ๐ฅ 2
๐ฅ ห 1
1.0 โข ๐ฅ 0 โข ๐ฅ 2 โ 1.0 โข ๐ฅ 1
๐ฅ ห 2
โ 0.02 โข ๐ฅ 0 โข ๐ฅ 2 + 1.0 โข ๐ฅ 1 โ 0.15 โข ๐ฅ 2
3 Fassoni2019 - Oncogenesis encompassing mutations and genetic instability
๐ฅ
ห
0
0.01 โ 0.01 โข ๐ฅ 0
๐ฅ ห 1
0.03 โข ๐ฅ 1
๐ฅ ห 2
โ 0.5 โข ๐ฅ 2 2 + 0.034 โข ๐ฅ 2
4 Zatorsky2006-p53-Model5
๐ฅ
ห
0
โ 3.7 โข ๐ฅ 0 โข ๐ฅ 1 + 2.0 โข ๐ฅ 0
๐ฅ ห 1
โ 0.9 โข ๐ฅ 1 + 1.1 โข ๐ฅ 2
๐ฅ ห 2
1.5 โข ๐ฅ 0 โ 1.1 โข ๐ฅ 2
5 Lenbury2001-InsulinKineticsModel-A
๐ฅ
ห
0
โ 0.1 โข ๐ฅ 0 โข ๐ฅ 2 + 0.2 โข ๐ฅ 1 โข ๐ฅ 2 + 0.1 โข ๐ฅ 2
๐ฅ ห 1
โ 0.01 โข ๐ฅ 0 + 0.01 + 0.01 / ๐ฅ 2
๐ฅ ห 2
โ 0.1 โข ๐ฅ 1 โข ๐ฅ 2 + 0.257 โข ๐ฅ 1 โ 0.1 โข ๐ฅ 2 2 + 0.331 โข ๐ฅ 2 โ 0.3187
6 Zatorsky2006-p53-Model1
๐ฅ
ห
0
โ 3.2 โข ๐ฅ 0 โข ๐ฅ 1 + 0.3
๐ฅ ห 1
โ 0.1 โข ๐ฅ 1 + 0.1 โข ๐ฅ 2
๐ฅ ห 2
0.4 โข ๐ฅ 0 โ 0.1 โข ๐ฅ 2
7 Smallbone2013 - Colon Crypt cycle - Version 1
๐ฅ
ห
0
โ 0.002207 โข ๐ฅ 0 2 โ 0.002207 โข ๐ฅ 0 โข ๐ฅ 1 โ 0.002207 โข ๐ฅ 0 โข ๐ฅ 2 + 0.1648 โข ๐ฅ 0
๐ฅ ห 1
โ 0.01312 โข ๐ฅ 0 2 โ 0.0216 โข ๐ฅ 0 โข ๐ฅ 1 โ 0.01312 โข ๐ฅ 0 โข ๐ฅ 2 + 1.574 โข ๐ฅ 0 โ 0.008477 โข ๐ฅ 1 2 โ 0.008477 โข ๐ฅ 1 โข ๐ฅ 2 + 0.5972 โข ๐ฅ 1
๐ฅ ห 2
โ 0.04052 โข ๐ฅ 0 โข ๐ฅ 1 โ 0.04052 โข ๐ฅ 1 2 โ 0.04052 โข ๐ฅ 1 โข ๐ฅ 2 + 4.863 โข ๐ฅ 1 โ 1.101 โข ๐ฅ 2
8 Cortes2019 - Optimality of the spontaneous prophage induction rate.
๐ฅ
ห
0
โ 0.99 โข ๐ฅ 0 2 / ( ๐ฅ 0 + ๐ฅ 1 ) โ 1.0 โข ๐ฅ 0 โข ๐ฅ 1 / ( ๐ฅ 0 + ๐ฅ 1 ) + 0.99 โข ๐ฅ 0
๐ฅ ห 1
โ 0.99 โข ๐ฅ 0 โข ๐ฅ 1 / ( ๐ฅ 0 + ๐ฅ 1 ) โ 1.0 โข ๐ฅ 1 2 / ( ๐ฅ 0 + ๐ฅ 1 ) + 1.0 โข ๐ฅ 1
๐ฅ ห 2
โ 0.001 โข ๐ฅ 2
9 Figueredo2013/2 - immunointeraction model with IL2
๐ฅ
ห
0
โ 1.0 โข ๐ฅ 0 โข ๐ฅ 1 / ( ๐ฅ 0 + 1.0 ) + 0.18 โข ๐ฅ 0
๐ฅ ห 1
0.05 โข ๐ฅ 0 + 0.124 โข ๐ฅ 1 โข ๐ฅ 2 / ( ๐ฅ 2 + 20.0 ) โ 0.03 โข ๐ฅ 1
๐ฅ ห 2
5.0 โข ๐ฅ 0 โข ๐ฅ 1 / ( ๐ฅ 0 + 10.0 ) โ 10.0 โข ๐ฅ 2
10 A mathematical model of cytotoxic and helper T cell interactions in a tumour microenvironment
๐ฅ
ห
0
โ 10.0 โข ๐ฅ 0 2 โ 2.075 โข ๐ฅ 0 โข ๐ฅ 2 + 10.0 โข ๐ฅ 0
๐ฅ ห 1
0.19 โข ๐ฅ 0 โข ๐ฅ 1 / ( ๐ฅ 0 2 + 0.0016 ) โ 1.0 โข ๐ฅ 1 + 0.5
๐ฅ ห 2
โ 2.075 โข ๐ฅ 0 โข ๐ฅ 2 + 1.0 โข ๐ฅ 1 โข ๐ฅ 2 โ 1.0 โข ๐ฅ 2 + 2.0
11 Munz2009 - Zombie SIZRC
๐ฅ
ห
0
โ 0.009 โข ๐ฅ 0 โข ๐ฅ 1 + 0.05
๐ฅ ห 1
0.004 โข ๐ฅ 0 โข ๐ฅ 1
๐ฅ ห 2
0.005 โข ๐ฅ 0 โข ๐ฅ 1 Table 9:Selected ODEBase dataset with variables ๐
3 . Report Issue Report Issue for Selection Generated by L A T E xml Instructions for reporting errors
We are continuing to improve HTML versions of papers, and your feedback helps enhance accessibility and mobile support. To report errors in the HTML that will help us improve conversion and rendering, choose any of the methods listed below:
Click the "Report Issue" button. Open a report feedback form via keyboard, use "Ctrl + ?". Make a text selection and click the "Report Issue for Selection" button near your cursor. You can use Alt+Y to toggle on and Alt+Shift+Y to toggle off accessible reporting links at each section.
Our team has already identified the following issues. We appreciate your time reviewing and reporting rendering errors we may not have found yet. Your efforts will help us improve the HTML versions for all readers, because disability should not be a barrier to accessing research. Thank you for your continued support in championing open access for all.
Have a free development cycle? Help support accessibility at arXiv! Our collaborators at LaTeXML maintain a list of packages that need conversion, and welcome developer contributions.
Xet Storage Details
- Size:
- 90.1 kB
- Xet hash:
- 830c478d7bbe3f5f4599df33f82b9dcdb815e01401609e9030b8a1c4cd883f03
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.