The present invention relates generally to the field of artificial intelligence-based modeling systems and more particularly relates to the learning of model parameters for non-linear systems of governing equations as described by differential equations.
Constitutive parameters in differential equations have been learned either analytically or computationally and in many cases the challenge in doing so can be significant. For example, learning parameters is conditioned on assumptions concerning the resolution and fidelity of the data, as well as the accuracy of the computationally generated solution to the differential equation, as well as the quality of the non-linear optimization algorithm. In the field of data-driven modeling, using artificial intelligence to learn model parameters for non-linear systems of differential equations (ordinary, partial, and stochastic) can be a significant challenge. Traditional methods often rely on numerical differential equation solvers, which can be computationally intensive and may not provide accurate estimates, especially in the presence of large levels of measurement noise. Furthermore, these methods may struggle with higher dimensional and stiff systems, leading to inefficiencies in both speed and accuracy.
Additionally, it is most common to use the so-called “strong” form representation of a model, which computes a weighted average value over time and space as opposed to a “weak” form which computes an average value over a region. Moreover, in addition to the challenges detailed above, the computational problems are further compounded when dealing with the Errors-In-Variables regression framework, which is used in statistical analysis of these models.
To perform parameter learning, the approximate solution must be computed for proposed parameter values and then valuation of the loss function involves computing a least squares difference between model and data. Learning of the parameters in a fully nonlinear model involves iteratively updating the proposed parameter values, leading to a minimization of the loss function. This is particularly relevant when dealing with models from various scientific, engineering, and medical fields, where accurate parameter estimation is crucial to ensure the validity and reliability of the models.
Therefore, there is a clear need for a more efficient and accurate method for estimating model parameters for non-linear systems of (ordinary, partial, stochastic) differential equations, particularly in the context of noisy data.
In accordance with one or more preferred embodiments of the present invention, a method is provided for estimating model parameters for non-linear systems of Differential Equations (DEs). The method involves receiving data from one or more measurement devices and applying one or more Weak-form Estimation of Nonlinear Dynamics (WENDy) methods to the received data. Note that in the creation of the weak form, a so-called “generalized function” is defined by the data and acts on the test function to convert the model system to the weak form. These WENDy methods are robust to instances of large measurement noise. The method also includes converting strong form representations of a model to weak forms, solving regression problems to perform parameter inference, and using an Errors-In-Variables frameworks such as scientific, engineering, and biomedical fields employing compartment-based ordinary differential equation, spatio-temporally-based partial differential equations.
In accordance with other embodiments, a system is provided for estimating model parameters for non-linear systems of Differential Equations (DEs). The system comprises a data receiver configured to receive data from one or more measurement devices, a processor configured to apply one or more Weak-form Estimation of Nonlinear Dynamics (WENDy) methods to the received data, and a memory configured to store one or more Errors-In-Variables frameworks and one or more iteratively reweighted least squares algorithms. The processor is also configured to convert strong form representations of a model to weak forms, solve regression problems to perform parameter inference, and create orthonormal test functions from Coo bump functions of varying support sizes. The system can handle both low dimensional systems with modest amounts of data and higher dimensional systems.
The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
The various embodiments of the present invention will hereinafter be described in conjunction with the appended drawings, wherein like designations denote like elements, and:
Accurate estimation of parameters for a given model is central to modern scientific discovery. It is particularly important in the modeling of biological systems which can involve both first principles-based and phenomenological models and for which measurement errors can be substantial, often in excess of 20%. The dominant methodologies for parameter inference are either not capable of handling realistic errors, or are computationally costly relying on forward solvers or Markov chain Monte Carlo methods. In this work, we propose an accurate, robust and efficient weak form-based approach to estimate parameters for parameter inference. We demonstrate that our “Weak form Estimation of Nonlinear Dynamics” (WENDy) method offers many advantages including high accuracy, robustness to substantial noise, and computational efficiency often up to several orders of magnitude over the existing methods.
In the remainder of this section, we provide an overview of modern parameter estimation methods in ODE systems, as well as a discussion of the literature that led to the WENDy idea. Section 2 contains the core weak-form estimation ideas as well as the WENDy algorithm itself. In Section 2.1, we introduce the idea of weak-form parameter estimation, including a simple algorithm to illustrate the idea. In Section 2.2, we describe the WENDy method in detail. We describe the Errors-In-Variables (EiV) framework, and derive a Taylor expansion of the residual which allows us to formulate the (in Section 2.2) Iteratively Reweighted Least Squares (IRLS) approach to inference. The EiV and IRLS modifications are important as they offers significant improvements to the Ordinary Least Squares approach. In Section 2.3, we present a strategy for computing an orthogonal set of test functions that facilitate a successful weak-form implementation. In Section 3 we illustrate the performance of WENDy using five common mathematical models from the biological sciences and in Section 4 we offer some concluding remarks.
A ubiquitous version of the parameter estimation problem in the biological sciences is
The ODE system in (2) is parameterized by w∈RJ, the vector of J true parameters which are to be estimated by ŵ. The solution to the equation is then compared (in a least squares sense) with data U∈R(M+1)×d that is sampled at M+1 timepoints t:={ti}i=0M. We note that in this work, we will restrict the differential equations to those with right sides that are linear combinations of the ƒj functions with coefficients wj, as in equation (2).
Conventionally, the standard approach for parameter estimation methodologies has been forward solver-based nonlinear least squares (FSNLS). In that framework: (i) a candidate parameter vector is proposed; (ii) the resulting equation is numerically solved on a computer; (iii) the output is compared (via least squares) to data; and (iv) then this process is repeated until a convergence criteria is met.
The FSNLS methodology is very well understood and its use is ubiquitous in the biological, medical, and bioengineering sciences. However, as models get larger and more realism is demanded of them, there remain several important challenges that do not have fully satisfying answers. For example, the accuracy of the solver can have a huge impact on parameter estimates in PDE models as well as ODE and DDE. There is no widespread convention on detection of this type of error and the conventional strategy would be to simply increase the solution accuracy (usually at significant computational cost) until the estimate stabilizes. Perhaps more importantly, the choice of the initial candidate parameter vector can have a huge impact upon the final estimate, given that nonlinear least squares cost functions frequently have multiple local minima in differential equations applications. There are several algorithms designed to deal with the multi-modality, such as particle swarm optimization and simulated annealing; however, all come at the cost of additional forward solves and unclear dependence on the hyperparameters used in the solver and optimization algorithms.
Given the above, it is reasonable to consider alternatives to fitting via comparing an approximate model solution with the measured data. A natural idea would be to avoid performing forward solves altogether via substituting the data directly into the model equation (2). The derivative could be approximated via differentiating a projection of the data onto, e.g., orthogonal polynomials, and the parameters could then be estimated by minimizing the norm of the residual of the equation (2)—i.e., via a gradient matching criterion. There have been similar ideas in the literature of chemical and aerospace engineering, which can be traced back even further. However, these methods are known to perform poorly in the presence of even modest noise.
To account for the noise in the measurements while estimating the parameters (and in some cases the state trajectories), researchers have proposed a variety of different non-solver-based methods. The most popular modern approaches involve denoising the measured state via Gaussian Processes and collocations projecting onto a polynomial or spline basis. For example, a Gaussian Process restricted to the manifold of solutions to an ODE may be used to infer both the parameters and the state using a Hamiltonian Markov chain Monte Carlo method. Similarly, a collocation-type method in which the solution is projected onto a spline basis may be considered. In a two-step procedure, both the basis weights and the unknown parameters are iteratively estimated. The minimization identifies the states and the parameters by penalizing poor faithfulness to the model equation (i.e., gradient matching) and deviations too far from the measured data. A similar strategy, based on local polynomial smoothing to first estimate the state and its derivative, compute derivatives of the smoothed solution, and then estimate the parameters may be considered. Improvements may be realized by using local polynomial regression instead of the pseudo-least squares estimator.
There are also a few approaches which focus on transforming the equations with operators that allow efficiently solving for the parameters. In particular, smoothing and derivative smoothing operators based on Fourier theory and Chebyshev operators may be considered. However, they have not proven to be as influential as the integral and weak form methods described in the next subsection.
Recent efforts by our group and others suggest that there is a considerable advantage in parameter estimation performance to be gained from using an integral-based transform of the model equations. The two main approaches are to: (i) use integral forms of the model equation; or (ii) convolve the equation with a compactly supported test function to obtain the so-called “weak form” of the equation. The weak form idea can be traced back to Laurent Schwartz's Theory of Distributions, which recasts the classical notion of a function acting on a point to one acting on a measurement structure or “test function”. In the context of differential equation models, Lax and Milgram pioneered the use of the weak form for relaxing smoothness requirements on unique solutions to parabolic PDE systems in Hilbert spaces. Since then, the weak form has been heavily used in studying solutions to PDEs as well as numerically solving for the solutions (e.g., the Finite Element Method), but not with the goal of directly estimating parameters.
The idea of weak-form based estimation has been repeatedly discovered over the years. Briefly, in 1954, a proto-weak-form parameter inference method, called the Equations Of Motion (EOM) method was created. Using this method, it was proposed to multiply the model equations by so-called method functions, i.e., what would now be called test functions. These test functions were based on sinn(vt) for different values of v and n. This method has alternatively been identified as the Modulating Function (MF) method. Both proposed and advocated for the use of polynomial test functions. The issue with these approaches (and indeed all subsequent developments based on these methods) is that the maximum power n is chosen to exactly match the number of derivatives needed to perform integration by parts (IBP). As now known, this choice means that these methods are not nearly as effective as they could be. As shown in the most preferred embodiments of the present invention, a critical step in obtaining robust and accurate parameter estimation is to use highly smooth test functions, e.g., to have n be substantially higher than the minimum needed by the IBP. This understanding led to the use of the Co bump functions in the most preferred embodiments of the present invention as disclosed herein (see Section 2.3).
In the relevant literature, there are several examples of using integral or weak-form equations that illustrate an integral-based approach in addition to efforts to use the integral form for parameter estimation. Concerning the weak form, several researchers have used it as a core part of their estimation methods. Unlike the most preferred embodiments of the present invention, however, either these approaches smooth the data before substitution into the model equation (which can lead to poor performance) or still require forward solves. As with the EOM and MF method described above, the test functions in these methods were also chosen with insufficient smoothness to yield the highly robust parameter estimates obtained from the most preferred embodiments of the present invention.
As the field of SINDy-based equation learning is built upon direct parameter estimation methods, there are also several relevant contributions from this literature. Previous efforts have shown that parameter estimation and learning an integral form of equations can be done in the presence of significant noise. Broadly speaking, however, the consensus has emerged that the weak form is more effective than a straightforward integral representation. In particular, several groups independently proposed weak form-based approaches. The weak form is now even implemented in the PySINDy code which is actively developed by the authors of the original SINDy papers. However, it should be noted that the Weak SINDy in PySINDy is based on an early weak form implementation.
While weak form methodology has been considered in the past, the most preferred embodiments of the present invention provide unique enhancements suitable for use for equation learning in a wide range of model structures and applications including: ODEs, PDEs, interacting particle systems of the first and second order, and online streaming. We have also studied and advanced the computational method itself. Among other contributions, we were the first to automate (with mathematical justification) test function hyperparameter specification, feature matrix rescaling (to ensure stable computations), and to filter high frequency noise [34]. Lastly we have also studied the theoretical convergence properties for WSINDy in the continuum data limit [36]. Among the results are a description of a broad class of models for which the asymptotic limit of continuum data can overcome any noise level to yield both an accurately learned equation and a correct parameter estimate (see for more information).
In this work, we assume that the exact form of a differential equation-based mathematical model is known, but that the precise values of constituent parameters are to be estimated using existing data. As the model equation is not being learned, this is different than the WSINDy methodology and, importantly, does not use sparse regression. We thus denote the method presented in this paper as the Weak-form Estimation of Nonlinear Dynamics (WENDy) method.
In Section 2.1, we start with an introduction to the idea of weak-form parameter estimation in a simple OLS setting. In Section 2.2 we describe the WENDy algorithm in detail, along with several strategies for improving the accuracy: in Section 2.3 we describe a strategy for optimal test function selection, and in Section 2.4 the strategy for improved iteration termination criteria.
We begin by considering a d-dimensional matrix form of (2), i.e., an ordinary differential equation system model
As the compact support of ϕ implies that ϕ(0)=ϕ(T)=0, this yields a transform of (3) into
This weak form of the equation allows us to define a novel methodology for estimating the entries in W.
Observations of states of this system are (in this paper) assumed to occur at a discrete set of M+1 timepoints {tm}m=0M with uniform stepsize Δt. The test functions are thus centered at a subsequence of K timepoints {tm
Approximating the integrals in (4) using a Newton-Cotes quadrature yields
Where
The Q matrix contains the quadrature weights on the diagonal. In this work we use the composite Trapezoidal rule 3 for which the matrix is
We defer full consideration of the integration error until Section 2.3 but note that in the case of a non-uniform timegrid, Q would simply be adapted with the correct stepsize and quadrature weights.
The core idea of the weak-form-based direct parameter estimation is to identify W as a least squares solution to
The ordinary least squares (OLS) solution to (6) is presented in Algorithm 1. We note that we have written the algorithm this way to promote clarity concerning the weak-form estimation idea. For actual implementation, we create a different Θi for each variable i=1 . . . , d and use regression for state i to solve for a vector ŵi of parameters (instead of a matrix of parameters W, which can contain values known to be zero). To increase computational efficiency, it is important to remove any redundancies and use sparse computations whenever possible.
The OLS solution has respectable performance in some cases, but in general there is a clear need for improvement upon OLS. In particular, it should be noted that (6) is not a standard least squares problem. The (likely noisy) observations of the state u appear on both sides of (5). Those skilled in the art will recognize this as an Errors in Variables (EiV) problem. While a full and rigorous analysis of the statistical properties of weak-form estimation is beyond the scope of this disclosure, several formal derivations aimed at improving the accuracy of parameter estimation are presented. These improvements are critical as the OLS approach is not generally considered to be reliably accurate. Accordingly, we define WENDy (in the next section) as a weak-form parameter estimation method which uses techniques that address the EiV challenges.
In this subsection, it should be acknowledged that the regression problem does not fit within the framework of ordinary least squares (see
Begin with the sampled data U:=u*+ε∈R(M+1)×d and vector of parameters to be identified w∈RJd. Use bolded variables to represent evaluation at the timegrid t, and use superscript * notation to denote quantities based on true (noise-free) parameter or states. The residual becomes
It is now possible to decompose the residual into several components
Where
Here, r0 is the residual without measurement noise or integration errors, and eint is the numerical integration error induced by the quadrature (and will be analyzed in Section 2.3).
Further considering the leftover terms eΘ−bε and taking a Taylor expansion around the data U
The first order matrix in the expansion (9) is
As mentioned above, it is assumed that all elements of ε are i.i.d. Gaussian, i.e., N(0,σ2) and thus to first order the residual is characterized by
In the case where w=w* and the integration error is negligible, (10) simplifies
It should be noted that in (11) (and in (10)), the covariance is dependent upon the parameter vector w. In the statistical inference literature, the Iteratively Reweighted Least Squares (IRLS) method offers a strategy to account for a parameter-dependent covariance by iterating between solving for w and updating the covariance matrix C. Furthermore, while the normality in (11) is approximate, the weighted least squares estimator has been shown to be consistent under fairly general conditions even without normality. In Algorithm 2 the WENDy method is presented, updating C(n) (at the n-th iteration step) in lines 7-8 and then the new parameters w(n+1) are computed in line 9 by weighted least squares.
The IRLS step in line 9 requires inverting C(n), which is done by computing its Cholesky factorization and then applying the inverse to G and b. Since this inversion may be unstable, it is desirable to allow for possible regularization of C(n) in line 8 via a convex combination between the analytical first-order covariance L(n)(L(n))T and the identity via the covariance relaxation parameter a. This regularization allows the user to interpolate between the OLS solution (α=1) and the unregularized IRLS solution (α=0). In this way WENDy extends and encapsulates Algorithm 1. However, in the numerical examples below, it is possible to simply set α=10−10 throughout, as the aforementioned instability was not an issue. Lastly, those skilled in the art will recognize that any iterative scheme needs a stopping criteria and this is further discussed in Section 2.4.
⊗ (ΦΘ(U))]
⊗ {dot over (Φ)}]
The outputs of Algorithm 2 include the estimated parameters ŵ as well as the covariance Ĉ of the response vector b such that approximately
A primary benefit of the most preferred embodiments of the present methodology disclosed herein is that the parameter covariance matrix S can be estimated from Ĉ using
This yields the variances of individual components of ŵalong diag (S) as well as the correlations between elements of ŵ in the off-diagonals of S. Here {circumflex over (σ)}2 is an estimate of the measurement variance σ2, which we compute by convolving each compartment of the data U with a high-order7 filter ƒ and taking the Frobenius norm of the resulting convolved data matrix ƒ*U. Throughout ƒ is set to be the centered finite difference weights of order 6 over 15 equally-spaced points (computed using [17]), so that ƒ has order 5. The filter ƒ is then normalized to have unit 2-norm. This yields a high-accuracy approximation of σ2 for underlying data u* that is locally well-approximated by polynomials up to degree 5.
When using WENDy for parameter estimation, a valid question concerns the choice of test function. This may be particularly challenging in the sparse data regime, where integration errors can easily affect parameter estimates. As previously noted, using higher order polynomials as test functions yielded more accuracy (up to machine precision). Inspired by this result and to render moot the question of what order polynomial is needed, it is considered desirable to develop a 2-step process for offline computation of highly efficient test functions, given a timegrid t.
The first step is to derive an estimator of the integration error that can be computed using the noisy data U and used to detect a minimal radius mt such that mt>mt leads to negligible integration error compared to the errors introduced by random noise. Inspired by wavelet decompositions; next, row-concatenate convolution matrices of test functions at different radii mt:=(2kmt; l={0, . . . , l−}). An SVD of this tall matrix yields an orthonormal test function matrix Φ, which maximally extracts information across different scales. It should be noted that in the later examples l−=3, which in many cases leads to a largest test function support covering half of the time domain.
To begin, consider a C∞ bump function
With the ψ in (13) we have discovered that the accuracy of the parameter estimates is relatively insensitive to a wide range of η values. Therefore, based on empirical investigation we arbitrarily choose η=9 in all examples and defer more extensive analysis to future work. In the rest of this section, we will describe the computation of mt and how to use ψ to construct Φ and Φ⋅.
In (9), it is clear that reducing the numerical integration errors eint will improve the estimate accuracy.
Considering the k-th element of
By expanding
into its Fourier Series we then have
Referring now to
It is now possible to derive a surrogate approximation of eint using the noisy data U to estimate this transition from integration error-dominated to noise error-dominated residuals. From the noisy data U on timegrid t∈RM, it is desirable to compute eint (u*,ϕk,M) by substituting U for u* and using the discrete Fourier transform (DFT), however the highest accessible mode is {circumflex over (F)}±M/2[ϕU]. On the other hand, it is possible to approximate eint (u*,ϕk,└M/s┘) from U, that is, the integration error over a coarsened timegrid (0, , 2
, . . . , └M/s┘
), where
=T/└M/s┘ and s>2 is a chosen coarsening factor. By introducing the truncated error formula
The result is
Statistically, under this additive noise model, êint (U,ϕk,└M/s┘,s) is an unbiased estimator of êint (u*,ϕk,└M/s┘,s), i.e.,
Picking radius mt as a changepoint of log(êrms), where êrms is the root-mean-squared integration error over test functions placed along the timeseries,
Having computed the minimal radius mt, it is possible to construct the test function matrices (Φ,Φ⋅) by orthonormalizing and truncating a concatenation of test function matrices with mt:=mt×(1,2,4,8). Letting Ψl be the convolution matrix for ψ(⋅;2lmtΔt), we compute the SVD of
Referring now to
The right singular vectors V then form an orthonormal basis for the set of test functions forming the rows of Ψ. Letting r be the rank of Ψ, we then truncate the SVD to rank K, where K is selected as the changepoint in the cumulative sum of the singular values (Σii)i=1r. Let
Having formed the test function matrices {Φ,Φ⋅}, the remaining unspecified process in Algorithm 2 is the stopping criteria SC. The iteration can stop in one of three ways: (1) the iterates reach a fixed point, (2) the number of iterates exceeds a specified limit, or (3) the residuals
Referring now to
With the fixed-point tolerance set to τFP=10−6, the S−W tolerance and starting point to τSW=10−4 and n0=10, and max_its=100.
The effectiveness of the most preferred embodiments of the present invention can be demonstrated by applying the methods to five ordinary differential equations canonical to biology and biochemical modeling. As demonstrated in the works mentioned in Section 1, it is known that the weak or integral formulations are advantageous, with previous works mostly advocating for a two-step process involving (i) pre-smoothing the data before (ii) solving for parameters using ordinary least squares. The most preferred embodiments of the present invention do not necessarily involve smoothing the data, and instead leverages the covariance structure introduced by the weak form to iteratively reduce errors in the ordinary least squares (OLS) weak-form estimation. Utilizing the covariance structure in this way not only reduces error, but reveals parameter uncertainties as demonstrated in Section 3.3.
Referring now to
We compare the WENDy solution to the weak-form ordinary least squares solution (described in Section 2 and denoted simply by OLS in this section) to forward solver-based nonlinear least squares (FSNLS). Comparison to OLS is important due to the growing use of weak formulations in joint equation learning/parameter estimation tasks, but often without smoothing or further variance reduction steps. In most cases WENDy reduces the OLS error by 60%-90% (see the bar plots in
In all cases below, it is possible to solve for approximate weights ŵ using Algorithm 2 over 100 independent trials of additive Gaussian noise with standard deviation σ=σNR∥vec (U*)∥rms for a range of noise ratios σNR. This specification of the variance implies that
As well as σNR, we vary the stepsize Δt (keeping the final time T fixed for each example), to demonstrate large and small sample behavior. For each example, a high-fidelity solution is obtained on a fine grid (512 timepoints for Logistic Growth, 1024 for all other examples), which is then subsampled by factors of 2 to obtain coarser datasets.
To evaluate the performance of WENDy, we record the relative coefficient error
The data Û is obtained by simulating forward the model using the learned coefficients ŵ from the exact initial conditions u(0) using the same Δt as the data. The RK45 algorithm is used for all forward simulations (unless otherwise specified) with relative and absolute tolerances of 10−12. Comparison with OLS solutions is displayed in bar graphs which give the drop in error from the OLS solution to the WENDy solution as a percentage of the error in the OLS solution.
The logistic growth model is the simplest nonlinear model for population growth, yet the u2 nonlinearity generates a bias that affects the OLS solution more strongly as noise increases.
Referring now to
The Lotka-Volterra model is generally a system of equations designed to capture predator-prey dynamics. Each term in the model is unbiased when evaluated at noisy data (under the i.i.d. assumption), so that the first-order residual expansion utilized in WENDy is highly accurate. The bottom right plot in
Referring now to
The Fitzhugh-Nagumo equations are considered a simplified model for an excitable neuron. The equations contain six fundamental terms with coefficients to be identified. The cubic nonlinearity implies that the first-order covariance expansion in WENDy becomes inaccurate at high levels of noise. Nevertheless,
Referring now to
The Hindmarsh-Rose model is typically used to emulate neuronal bursting and features 10 fundamental parameters which span 4 orders of magnitude. Bursting behavior is observed in the first two solution components, while the third component represents slow neuronal adaptation with dynamics that are two orders of magnitude smaller in amplitude. Bursting produces steep gradients which render the dynamics numerically discontinuous at M=128 timepoints, while at M=256 there is at most one data point between peaks and troughs of bursts (see
Referring now to
The PTB model is a five-compartment protein transduction model identified in as a mechanism in the signaling cascade of epidermal growth factor (EGF). It was used in to compare between four other models, and has since served as a benchmark for parameter estimation studies in biochemistry. The nonlinearites are quadratic and sigmoidal, the latter category producing nontrivial transformations of the additive noise. WENDy estimates the 11 parameters with reasonable accuracy when 256 or more timepoints are available (see
In addition to the examples presented above, the most preferred embodiments of the present invention may be used to inform the user about uncertainties in the parameter estimates.
It is now possible to briefly compare WENDy and forward solver-based nonlinear least squares (FSNLS) using walltime and relative coefficient error E2 as criteria. For nonlinear least-squares one must specify the initial conditions for the ODE solve (IC), a simulation method (SM), and an initial guess for the parameters (w(0)). Additionally, stopping tolerances for the optimization method must be specified (Levenberg-Marquardt is used throughout). Optimal choices for each of these hyperparameters is an ongoing area of research. For this example, an optimized FSNLS is modeled in ways that are unrealistic in practice in order to demonstrate the advantages of the most preferred embodiments of the present invention even when FSNLS is performing somewhat optimally in both walltime and accuracy. The relevant hyperparameter selections are collected in Table 2 and discussed below.
To remove some sources of error from FSNLS, the true initial conditions u(0) are used throughout, noting that these would not be available in practice. For the simulation method, state-of-the-art ODE solvers are used for each problem, namely for the stiff differential equations Fitzhugh-Nagumo and Hindmarsh-Rose we use MATLAB's ode15s, while for Lotka-Volterra and PTB ode45 is selected. In this way FSNLS is optimized for speed in each problem. The relative and absolute tolerances of the solvers are fixed at 10−6 in order to prevent numerical errors from affecting results without asking for excessive computations. In practice, the ODE tolerance, as well as the solver, must be optimized to depend on the noise in the data, and the relation between simulation errors and parameters errors in FSNLS is an on-going area of research.
Caption for
Caption for
Caption for
Due to the non-convexity of the loss function in FSNLS, choosing a good initial guess w(0) for the parameters w* is crucial. For comparison, two strategies are employed. The first strategy (simply labeled FSNLS in
For the second initial guess strategy, let w(0)=ŵ, the output from WENDy (labeled WENDy-FSNLS in
Caption for
For Hindmarsh-Rose, even with high-resolution data and low noise (bottom left plot of
Finally, the aggregate performance of WENDy, WENDy-FSNLS, and FSNLS is reported in
Caption for
Caption for
Caption for
Caption for
In this disclosure, it is demonstrated that the Weak-form Estimation of Nonlinear Dynamics (WENDy) method of the present invention are well-suited for directly estimating model parameters, without relying on forward solvers. The essential feature of the most preferred methods involve converting the strong form representation of a model to its weak form and then substituting in the data and solving a regression problem for the parameters. The method is robust to substantial amounts of noise, and in particular to levels frequently seen in biological experiments.
As mentioned above, the idea of substituting data into the weak form of an equation followed by a least squares solve for the parameters has existed since at least the mid 1950's. However, FSNLS-based methods have proven highly successful and are ubiquitous in the parameter estimation literature and software. The disadvantage of FSNLS is that fitting using repeated forward solves comes at a substantial computational cost and with unclear dependence on the initial guess and hyperparameters (in both the solver and the optimizer). Several researchers over the years have created direct parameter estimation methods (that do not rely on forward solves), but they have historically included some sort of data smoothing step. The primary issue with this is that projecting the data onto a spline basis (for example) represents the data using a basis which does not solve the original equation. Importantly, that error propagates to the error in the parameter estimates. However, we note that the WENDy framework introduced here is able to encapsulate previous works that incorporate smoothing, namely by including the smoothing operator in the covariance matrix C.
The conversion to the weak form is essentially a weighted integral transform of the equation. As there is no projection onto a non-solution based function basis, the weak-form approach bypasses the need to estimate the true solution to directly estimate the parameters.
One of the most salient and unique aspects of the present invention is the teaching that the weak-form-based direct parameter estimation methods disclosed herein significant advantages over traditional FSNLS-based methods. In almost all the examples disclosed herein and, in particular for larger dimensional systems with high noise, the unique methods of the present invention are faster and more accurate by orders of magnitude. In rare cases where an FSNLS-based approach yields higher accuracy, the methods of the present invention may still be used as an efficient method to identify a good initial guess for parameters.
It is to be understood that although aspects of the present specification are highlighted by referring to one or more specific embodiments, those skilled in the art will readily appreciate that these disclosed embodiments are only illustrative of the principles of the subject matter disclosed herein. For example, although the disclosure refers to the various preferred embodiments of the present invention primarily in conjunction with certain models for specific applications, those skilled in the art will recognize that the various embodiments of the present invention are suitable for use in conjunction with other models and applications where more accurate estimation of variables and parameters for various models is desirable.
Therefore, it should be understood that the disclosed subject matter is in no way limited to a particular methodology, protocol, and/or material, etc., described herein. As such, various modifications or changes to or alternative configurations of the disclosed subject matter can be made in accordance with the teachings herein without departing from the spirit of the present specification. Further, the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to limit the scope of the present disclosure, which is defined solely by the claims. Accordingly, embodiments of the present disclosure are not limited to those precisely as shown and described.
Unless otherwise indicated, all numbers expressing a characteristic, item, quantity, parameter, property, term, and so forth used in the present specification and claims are to be understood as being modified in all instances by the term “about.” As used herein, the term “about” means that the characteristic, item, quantity, parameter, property, or term so qualified encompasses a range of plus or minus ten percent above and below the value of the stated characteristic, item, quantity, parameter, property, or term. Accordingly, unless indicated to the contrary, the numerical parameters set forth in the specification and attached claims are approximations that may vary. At the very least, and not as an attempt to limit the application of the doctrine of equivalents to the scope of the claims, each numerical indication should at least be construed in light of the number of reported significant digits and by applying ordinary rounding techniques.
Notwithstanding that the numerical ranges and values setting forth the broad scope of the disclosure are approximations, the numerical ranges and values set forth in the specific examples are reported as precisely as possible. Any numerical range or value, however, inherently contains certain errors necessarily resulting from the standard deviation found in their respective testing measurements. Recitation of numerical ranges of values herein is merely intended to serve as a shorthand method of referring individually to each separate numerical value falling within the range. Unless otherwise indicated herein, each individual value of a numerical range is incorporated into the present specification as if it were individually recited herein.
The terms “a,” “an,” “the” and similar references used in the context of describing the disclosed embodiments (especially in the context of the following claims) are to be construed to cover both the singular and the plural, unless otherwise indicated herein or clearly contradicted by context. All methods described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. The use of any and all examples, or exemplary language (e.g., “such as”) provided herein is intended merely to better illuminate the present disclosure and does not pose a limitation on the scope of the embodiments otherwise claimed. No language in the present specification should be construed as indicating any non-claimed element essential to the practice of the disclosed embodiments.
This invention was made with government support under grant number R01 GM126559 awarded by the National Institutes of Health, grant numbers MCB2054085 and DEB2109774 awarded by the National Science Foundation, grant number DE-SC0023346 awarded by the US Department of Energy, DEB2109774 awarded by the National Science Foundation, and grant number 2019-67014-29919 awarded by the US Department of Agriculture. The government has certain rights in the invention.
| Number | Date | Country | |
|---|---|---|---|
| 63621086 | Jan 2024 | US |