Buckets:
Title: A Hybrid Two-level MCMC Framework to Accelerate Posterior Mean Estimation with Deep Learning Surrogates for Bayesian Inverse Problems
URL Source: https://arxiv.org/html/2307.01463
Markdown Content: Abstract 1Introduction 2A new approach to accelerate Bayesian inversion 3Numerical experiments 4Conclusion References A Hybrid Two-level MCMC Framework to Accelerate Posterior Mean Estimation with Deep Learning Surrogates for Bayesian Inverse Problems Juntao Yang Jeff Adie Simon See Adriano Gualandi Gianmarco Mengaldo Abstract
Bayesian inverse problems arise in various scientific and engineering domains, and solving them can be computationally demanding. This is especially the case for problems governed by partial differential equations, where the repeated evaluation of the forward operator is extremely expensive. Recent advances in Deep Learning (DL)-based surrogate models have shown promising potential to accelerate the solution of such problems. However, despite their ability to learn from complex data, DL-based surrogate models generally cannot match the accuracy of high-fidelity numerical models, which limits their practical applicability. We propose a novel hybrid two-level Markov Chain Monte Carlo (MCMC) method that combines the strengths of DL-based surrogate models and high-fidelity numerical solvers to compute the posterior mean of Quantities of Interest (QoI) in Bayesian inverse problems governed by partial differential equations. The intuition is to leverage the evaluation speed of a DL-based surrogate model as the base chain, and correct its errors using a limited number of high-fidelity numerical model evaluations in a correction chain; hence its name hybrid two-level MCMC method. Through a detailed theoretical analysis, we show that our approach can achieve the same accuracy as a pure numerical MCMC method while requiring only a small fraction of the computational cost. The theoretical analysis is further supported by several numerical experiments, namely a Poisson, a non-linear reaction-diffusion, and a Navier-Stokes equation. The proposed hybrid framework can be generalized to other approaches such as the ensemble Kalman filter and sequential Monte Carlo methods.
keywords: Markov Chain Monte Carlo , Deep Learning , Bayesian Inverse Problems β journal: Journal of Computational Physics \affiliation
[nvidia]organization=NVIDIA AI Technology Center, addressline=07-03 Suntec Tower Three, 8 Temasek Blvd, country=Singapore \affiliation[cambridge]organization=University of Cambridge, addressline=Downing St., Cambridge CB2 3EQ, country=United Kingdom \affiliation[INGV]organization=Istituto Nazionale di Geofisica e Vulcanologia, addressline=Via di Vigna Murata, 605, 00143 Roma RM, country=Italy \affiliation[nus]organization=National University of Singapore, addressline=21 Lower Kent Ridge Rd, country=Singapore
1Introduction
Inverse problems arise in various fields of applied science, including design optimization in engineering, seismic inversion in geophysics, and data assimilation in weather forecasting [MR2102218]. The behavior of these systems is described by a mathematical model that frequently consists of a system of partial differential equations that depends on a set of inputs and parameters. Inverse problems involve determining the inputs or parameters of the mathematical model based on observations or partial observations of the model solution. The mathematical model, in the context of inverse problems, is also known as the forward problem, and it is typically expressed as
π¦
π’ β ( π§ ) ,
(1)
where π’ is the forward operator (also referred to as the forward map), π§ represents the inputs or parameters, and π¦ are the observed data defined in Equation (1).
The objective of an inverse problem is to identify the inputs or parameters π§ , or some Quantities of Interest (QoI) that depends on π§ , denoted as π β ( π§ ) . This can be, for instance, the permeability field of a Darcy flowβs subsurface model, or the initial condition of a Navier Stokes equation. Optimization techniques, such as least squares optimization, are commonly employed to solve inverse problems [MR2130010, MR3285819]. However, inverse problems are often ill-posed, meaning they may lack uniqueness, stability, or the existence of a solution.
To address the challenges related to ill-posedness, Tikhonov regularization is frequently used [MR3285819, MR2102218], where the inverse problem is solved as an optimization problem
argmin π§ β π β ( 1 2 β β π¦ β π’ β ( π§ ) β π 2 + 1 2 β β π§ β π 0 β π 2 ) ,
(2)
with norms β₯ β β₯ π and β₯ β β₯ π defined on two Banach spaces, namely π and π representing the data and model space, respectively. Although the incorporation of Tikhonov regularization might initially seem arbitrary, it can be explicitly interpreted from a Bayesian perspective as a prior distribution. This connection bridges the optimization approach with the probabilistic Bayesian framework, where data are considered as observations subject to noise together with a prior belief on the parameters or inputs π§ , namely π 0 in Equation 2. In this context, a noise term π is added to the forward operator defined in Equation (1) to account for observational noise,
π¦
π’ β ( π§ ) + π .
(3)
From Equation (3), we can write a posterior distribution for the model parameters given the observations as follows
πΎ π¦
π ( π§ | π¦ )
π ( π¦ | π§ ) π ( π§ ) π β ( π¦ ) ,
(4)
where π β ( π¦ ) is the evidence of the data, π β ( π§ ) is the prior probability about the model parameters, and π ( π¦ | π§ ) is the likelihood of the given observations. Rather than solving for the full posterior distribution, it is often convenient to solve a maximization problem just for the numerator of the right hand side. Under the assumption that the prior distribution of π§ as well as the distribution of π are Guassian (the latter with zero mean), this is equivalent to maximizing
πΎ π¦ β π ( π¦ | π§ ) π ( π§ ) β exp ( β 1 2 β₯ π¦ β π’ ( π§ ) β₯ π 2 β 1 2 β₯ π§ β π 0 β₯ π 2 ) ,
(5)
where norms β₯ β β₯ π and β₯ β β₯ π are defined on the covariance of the prior ( π ) and of the noise ( π ). Finding the maximum a posterior (MAP) of the distribution in Equation (5), leads to the same optimization problem as Equation (2).
In this work, we approach inverse problems, such as the one in Equation (3), leveraging Bayesβ rule (Equation 4). In this Bayesian context, we can calculate the posterior distribution, which reflects the updated beliefs about the unknowns after observing the data, even under more general assumptions than Gaussianity of π and of the prior distribution of π§ . Bayesβ theorem is often expressed formally with measure-theoretic terms from the mathematical framework of the Radon-Nikodym theorem, such that probability masses or densities over real numbers can be extended to probability measures over any arbitrary sets [MR2767184]. In this framework, the posterior measure and the prior measure are related through the Radon-Nikodym derivative [MR2652785].
Bayesian inverse problems can be finite- [MR3285819], or infinite-dimensional [MR2652785]. The former usually arises in the context of parameter estimation, whereby a finite set of parameters is of interest. The latter instead arises in the context of full-field inversion of partial differential equations (PDE) problems, where infinite-dimensional functions are of interest. In our paper, we adopt the Bayes theorem in measure-theoretic terms, which is compatible with the infinite-dimensional setup.
Solving Bayesian inverse problems typically leads to the repeated solution of the forward problem in equation (3). For example, to solve PDE-based (i.e. infinite-dimensional) Bayesian inverse problems, it is necessary to discretize the continuous PDE problem via a suitable numerical method, such as the finite element method [CANTWELL2015205] or the finite volume method [MR1925043]. This often leads to a high-dimensional linear system of equations, that are extremely expensive computationally. These computationally expensive high-dimensional linear systems, in turn, need to be solved several times to approximate the posterior distribution, making the problem intractable due to the curse of dimensionality.
Recent developments in deep learning (DL) have provided a possible pathway to accelerate the solution of Bayesian inverse problems. In particular, DL-based models (i.e., deep neural networks) can be used as surrogate models to substitute the computationally expensive high-dimensional linear system of equations that arise when numerically discretizing the continuous PDE problem. Two of the critical advantages of DL-based surrogates are their fast differentiability (thanks to automatic differentiation) and fast evaluation. The first feature makes DL-based surrogates an excellent candidate for solving Bayesian inverse problems using deterministic methods, such as variational methods; see, e.g. [maulik2022efficient]. Variational methods typically lead to finding the maximum a posteriori (MAP) with optimization techniques. Gradient-based optimization such as gradient descent and L-BFGS is a family of the most used optimization techniques for variational methods. Gradients can be easily computed by automatic differentiation from a differentiable DL-based surrogate model, which makes the DL-based surrogate model a great fit. The second feature makes them an excellent candidate for sampling-based statistical methods, such as Markov Chain Monte Carlo and ensemble Kalman filter. These sampling techniques approximate the posterior distribution through Monte Carlo samples, which typically converge at a slow rate of 1 / π ( π being the number of samples). In this case, the fast evaluation speed of DL-based surrogates can be utilized to dramatically accelerate sampling procedures.
In the literature, several works explored the use of DL-based models for inverse problems. For example, physics-informed neural networks (PINNs) have demonstrated their ability to solve parameterized PDEs; feature that can be used for finite-dimensional inverse problems, such as design optimization [MR3881695, tay2025optimization]. It has been shown that neural operators can learn linear and non-linear mappings between function spaces [LU2022114778]; hence, they are a promising category of DL-based surrogates for infinite-dimensional inverse problems (e.g. [DBLP:conf/iclr/LiKALBSA21]). Indeed, the fast evaluation properties of neural operators make them extremely competitive for high-dimensional problems with respect to more traditional surrogate models, such as generalized polynomial chaos and Gaussian processes [marzouk2009stochastic, MR3084684]. These recent advances have made the solution of otherwise intractable inverse problems a real possibility.
However, despite the advantages of DL-based surrogate models for inverse problems, there are still some key areas that need to be addressed. Namely, a complete mathematical framework to estimate the error bounds of a deep neural network model is still missing. The expected error bounds (also referred interchangeably as error estimates) consist of three components: approximation error, optimization error, and generalization error [jin2020quantifying]. While universal approximation theorems exist [LESHNO1993861, chen1995universal], together with the expressivity analysis of neural networks (that depends on the number of layers and nodes provided) [petersen2018optimal, aadebffb88b448c89c654fcdda52c02b], these only address the approximation error. A large body of literature attempts to address the optimization error by investigating the landscape of non-convex loss functions as well as the optimization process by stochastic gradient [jin2020quantifying, pmlr-v97-allen-zhu19a, du2019gradient, 10.1093/imanum/drz055]. Some works also attempt to quantify the generalization error [jin2020quantifying]. However, a general theory is still lacking as most existing analyses make several simplifying assumptions that do not hold for practical problems.
Because of this lack of theoretical error framework, DL-based models are commonly treated as a black-box, and the expected error bounds are empirically estimated with a test dataset, noting that increasing test accuracy of DL models often requires exponentially more data (property known as the power law) [hestness2017deep, 10.1073/pnas.2311878121].
In the context of Bayesian inverse problems, the lack of a rigorous theoretical framework on DL modelsβ error bounds hinders the adoption of DL surrogates in critical applications, where a desired error estimated a priori may be required. In fact, in Bayesian inverse problems, a naive replacement of a high-fidelity numerical PDE solver with a DL-based surrogate will lead to propagation of the error of the surrogate to the posterior distribution [CAO2023112104]. For instance, in the context of MCMC methods, that sample the posterior distribution with an ergodic Markov Chain generated by a given algorithm (e.g., the Metropolis-Hasting algorithm [brooks2011handbook]), the estimation error of the posterior mean on the QoI depends on the surrogate error shown in Section 2.2.2.
Therefore, in order to make the DL-based surrogate model practically useful in solving Bayesian inverse problems, e.g., using MCMC methods, the posterior error induced by the DL surrogate needs to be contained to a given a priori threshold, an aspect that is still lacking and that represents a key gap in the literature.
In this work, we focus on MCMC methods that are among the most widely adopted methods for solving Bayesian inverse problems, given their ability to handle high-dimensional problems and their embedded uncertainty estimates. More specifically, we focus on problems that aim to compute statistical properties such as the posterior mean and variance of some QoI. However, MCMC methods have a critical issue: they are extremely expensive computationally. To address this issue, we propose a new MCMC approach to estimate the posterior mean in Bayesian inverse problems that we named two-level hybrid MCMC approach. The new method leverages the fast evaluation properties of DL surrogates to accelerate the MCMC method, while also using a high-fidelity numerical model for accuracy. The latter aspect allows for a priori theoretical error estimates of the posterior mean of the QoI which can be controlled by the choice of the high-fidelity numerical model. This is typically not readily available when only using DL surrogates.
Our method draws inspiration from numerical multilevel MCMC methods [doi:10.1137/130915005]. Generally speaking, there are two approaches to improve the computational cost of MCMC via multilevel methods.
The first approach uses coarser resolution numerical models as filters to pre-screen the proposed sample before going to the acceptance/rejection step in the Metropolis-Hasting algorithm with expensive high-resolution numerical models [MR4537564, Christen01122005, efendiev2005efficient, cui2011bayesian, MR2231730]. Hybrid MCMC algorithms were also proposed within this context, using traditional surrogate models (including generalized polynomial chaos and Gaussian processes) as the pre-screening filters [laloy2013efficient, reddy2024accelerating]. This first approach improves the sampling efficiency of the Metropolis-Hasting algorithm by improving acceptance rates at the finer resolution level. However, the number of effective samples required remains unchanged to reach a target error π β€ πΆ β π π β π β π β π β 1 / 2 , where πΆ is a constant and π π β π β π β π is the number of effective samples at fine resolution. For an elliptic partial differential equation-governed multiscale Bayesian inverse problem, the overall computational complexity of such approach remains at best πͺ β ( π β π β 2 ) where π is the desired approximation error for the posterior mean of QoI and π is the dimension of the problem [MR3084684].
The second approach is based on the idea of telescoping sum, upon which our method is based. The telescoping sum technique was initially proposed for the multilevel Monte Carlo [MR2436856], then it was extended to the MCMC method with some modifications [doi:10.1137/130915005, MR3084684, MR4523340]. Instead of focusing on sampling efficiency, this technique exploits the fact that the variance of the difference between solutions of two resolution levels πΏ and πΏ β 1 in a PDE-constrained Bayesian inverse problem typically decreases with larger πΏ , where πΏ is the level of mesh refinement with the mesh size β of scale πͺ β ( 2 β πΏ ) . This allows an efficient multilevel approach to achieve an accurate posterior mean approximation by the telescoping sum technique, where less expensive high resolution samples are needed when πΏ is large thanks to the smaller variance. In certain problems, such as the Bayesian inverse problem with elliptic PDEs with bi-hierarchical setup, the computational complexity can be reduced to πͺ β ( π β π ) [MR3084684].
Despite the attempts in [laloy2013efficient, reddy2024accelerating] to build hybrid MCMC methods using surrogate model in delayed-acceptance-like MCMC algorithms, there have been no attempts to build hybrid MCMC methods using surrogate models in the telescopic approach. The method we propose in this paper fills this gap.
We note that in the telescoping sum approach, each MCMC chain runs independently at different levels. As a consequence, there is no restriction on the type of MCMC algorithm that can be used to accelerate the sampling efficiency of each MCMC chain, and several potential methods can be used, including the Delayed Acceptance algorithm, the Preconditioned CrankβNicolson algorithm, and the Stochastic Newton MCMC method, among others [doi:10.1137/110845598, cotter2013mcmc].
With the telescoping sum technique, our two-level hybrid MCMC approach (also referred simply to as Hybrid MCMC) uses a DL-based surrogate model to obtain a base MCMC chain with a large number of samples (leveraging the fast evaluation speed of the DL surrogate). A short correction MCMC chain is then generated to sample the differences between the high-fidelity numerical model and the DL-based surrogate model. This is done to correct for the bias introduced by the surrogate as shown in Fig 1. Despite our focus on the fast developing deep learning-based surrogate models, the hybrid two level MCMC structure is generally applicable to all kinds of surrogate models including Polynomial Chaos, Gaussian Process Regressions and etc.
Figure 1:Hybrid two-level MCMC
We provide a detailed theoretical analysis, showing that the new method has the same a priori error bound πͺ β ( β ) as a plain, i.e. single chain, MCMC method that uses a high-fidelity numerical model, discretized at a known mesh size equal to β . However, our method requires a small fraction of the computational cost necessary to run a plain MCMC chain with a high-fidelity numerical model. Despite the computational advantages, we shall mention that the proposed approach is limited to the computation of the posterior mean of QoI, as it is not possible to generate a histogram as can be commonly done when using single-chain MCMC algorithms. Yet, several useful statistical quantities, such as variance, cumulative distribution function, a quantile, and the associated conditional value-at-risk, may still be estimated with techniques such as those mentioned in [doi:10.1137/17M1135566].
We complement the theoretical findings with numerical experiments on an elliptic, a reaction-diffusion, and a fluid dynamics problem. The numerical results support the theoretical findings and highlight how the new method provides a lightweight DL-based surrogate based alternative to existing MCMC approaches, with rigorously defined a priori error bounds. The latter aspect closes the gap in the literature regarding the lack of rigorous error bounds when using DL-based surrogates, and constitutes an important milestone for the fast solution of Bayesian inverse problems via deep learning. The rest of the paper is organized as follows. Section 2, introduces the Bayesian inverse problem setup (Section 2.1), the approximation of the forward problem (Section 2.2), and the new hybrid two-level MCMC for a uniform prior (Section 2.3), noting that the Gaussian prior case is presented in A. Section 3 shows the numerical experiments that validate the theoretical estimates provided in Section 2. Section 4 draws some closing remarks, including limitations and future work.
2A new approach to accelerate Bayesian inversion 2.1Bayesian inverse problem setup
To present our new approach, we first introduce the theoretical background of Bayesian inverse problems. We consider inverse problems governed by a forward mathematical model as the one defined in equation (3), where the underlying system is constituted of PDEs. More formally, the PDE-based forward model predicts the states π’ provided the inputs/parameters π³
{ π§ 1 , π§ 2 , β¦ , π§ π } . In order to introduce the problem setup based on Equation (3), we need to define the inputs/parameters π³ , the forward operator (or forward map) π’ β ( π³ ) , and the observational noise π .
We start by defining the inputs/parameters π³ . These represent a finite number of constants or functions within the governing equations, or the coefficients associated with the spectral expansion of a random field defining the initial conditions or forcing terms. For example, π³ can be the LamΓ© constants of the material in the elasticity equation of solid mechanics [MR2322235], the coefficient of the KarhunenβLoΓ¨ve expansion of the porosity random field in the subsurface flow model [MR2421969], or the initial condition and random forcing in the Navier-Stokes equations [MR2558668]. In many practical applications, the following truncated KarhunenβLoΓ¨ve expansion of a random field πΎ is commonly used,
πΎ β ( π§ )
πΎ Β― + β π
1 π π§ π β π π ,
(6)
where πΎ Β― , π π are functions in πΏ β β ( π· ) , where π· is the physical domain. For simplicity, hereafter we name π³ as the parameters of the forward problem, without lacking generality.
In the context of Bayesian inverse problems, we need to define the prior probability distribution for the parameters π³ . To this end, we consider a uniform prior, where π§ π in π³ is uniformly distributed within [ β 1 , 1 ] . By denoting β¬ the Borel π -algebra, we define a measurable space ( π , Ξ ) , where Ξ is a π -algebra on π , defined as Ξ
β¨ π
1 π β¬ β ( [ β 1 , 1 ] ) , and π
[ β 1 , 1 ] π is the parameter space. Together with the prior measure πΎ
β¨ π
1 π π β π§ π 2 on the measurable space ( π , Ξ ) , we have the complete probability space ( π , Ξ , πΎ ) .
With the prior measure of π³ defined, we focus on the forward operator (or map) π’ . Within the framework just introduced, the forward operator can be written as follows,
π’ : π β β π β π³ β π ; π’ ( π³ )
( β± 1 ( π’ ( π³ ) ) , β± 2 ( π’ ( π³ ) ) , β¦ , β± π ( π’ ( π³ ) ) ) ,
(7)
where π’ β ( π³ ) β π is the state solution of the forward PDEs which depends on the input π³ , and β± π β ( β ) , π
1 , 2 , β¦ β π are π continuous bounded linear functionals. π is a suitable vector space, e.g. a Sobolev space over π· , which depends on the specific physical problem. β± is included in the forward operator to better reflect real-world problems, where real-world observations are usually discrete and sparse while the state solutions of a PDE system are typically continuous functions. For example, in the context of weather data assimilation, β± is known as the observation operator [skamarock2019description]. In order to frame our Bayesian problem and guarantee the existence of the posterior, we need to formulate a key assumption on the forward operator π’ .
Assumption 2.1.
The forward operator π’ β ( π§ ) : π β β π is a continuous map from the measurable space ( π , Ξ ) to ( β π , β¬ β ( β π ) ) .
Assumption 2.1 is valid for most PDE-constrained systems, and it leads to the existence of the posterior in Bayesian inverse problems. Proofs of Assumption 2.1 for elliptic and parabolic equations can be found in [MR3084684, MR4246090], while the proofs for the elasticity and Navier-Stokes equations can be found in [MR2652785, MR2558668].
We finally define the observational noise, π . We assume it to be Gaussian and independent of the parameters π³ . Therefore, π is a random variable with values in β π and it follows a normal distribution π© β ( 0 , Ξ£ ) , where Ξ£ is a known π Γ π symmetric positive covariance matrix.
Having defined the parameters π³ , the forward operator π’ , and the observational noise π , along with the measurable space ( π , Ξ ) , and the prior distribution of the parameters πΎ , we now show the existence of the posterior, which we denote by πΎ π¦ . This is possible thanks to Assumption 2.1 that leads to πΎ π¦ being absolutely continuous with respect to the prior πΎ . The posterior probability measure is defined through the Radon-Nikodym derivative
π β πΎ π¦ π β πΎ β exp β‘ ( β Ξ¦ β ( π§ ; π¦ ) ) ,
(8)
where Ξ¦ is known as the potential function, for example with Gaussian noise assumption, Ξ¦ β ( π§ , π¦ )
1 2 β β π¦ β π’ β ( π§ ) β Ξ£ 2 . Detailed proof of equation (8) can be found in [MR2652785].
The last step to fully setup the framework is to show that the posterior measure is well-posed. This can be achieved following the results in [MR2652785, MR2558668, MR4069815], that detail how the posterior measure is Lipschitz continuous with respect to the data under a certain distance metric. Specifically, for every π > 0 and π¦ , π¦ β² β β π with β π¦ β Ξ£ , β π¦ β² β Ξ£ β€ π , there exists πΆ
πΆ β ( π )
0 such that
π π» β ( πΎ β² , πΎ β²β² )
( 1 2 β β« π ( π β πΎ β² π β πΎ β π β πΎ β²β² π β πΎ ) 2 β π πΎ ) 1 2 β€ πΆ β ( π ) β β π¦ β π¦ β² β Ξ£ ,
(9)
where πΎ β² and πΎ β²β² are two measures on π , which are absolutely continuous with respect to the measure πΎ , and where we chose as a distance metric the Hellinger distance π π» . The latter was chosen to facilitate various proofs related to our new hybrid two-level MCMC approach, leading to Theorem 2.2, in Section 2.3.
We note that the setup considered uses a uniform prior for the sake of simplicity. However, we can also work with a Gaussian setup, e.g. π
β π , Ξ
β¨ π
1 π β¬ β ( β ) , and πΎ
β¨ π
1 π π© β ( 0 , 1 ) .
2.2Approximation of the forward problem
A particularly expensive task in the Bayesian inverse problem setup introduced in Section 2.1 is the solution of the forward problem, especially when the forward operator π’ is constituted of PDEs. We distinguish two cases: (i) when the PDEs are approximated and solved via traditional numerical methods (e.g., FEM or others), also referred to as high-fidelity numerical models/solvers, and (ii) when the PDEs are solved via DL surrogates. We detail these two cases next, where, leveraging the results highlighted in Section 2.1, we derive theoretical error estimates for each case, and make some observations on the computational costs.
2.2.1Traditional approximation methods
The first case considered uses traditional numerical methods, such as FEM [parolini2005mathematical], SEM [mengaldo2021industry], or FVM [jameson1977finite], to discretize the PDE system. These numerical approximations lead to large linear systems of equations that are extremely computationally expensive, rendering the solution of forward problems impractical in the context of sampling-based statistical techniques, such as MCMC. Yet, they provide well-defined theoretical error estimates, that typically depend on how fine the discretization (i.e., the tessellation of the computational domain via elements or grid points, also known as mesh) is. This property is particularly useful in the context of inverse problems, because it allows practitioners to have a clear understanding of the errors incurred within their solution framework.
In particular, when considering any of the numerical methods above, we can define a priori estimates on the error we might expect for a certain discretization level β . The latter is an integer value that specifies a characteristic, β , that represents the dimensions of the elements (or spacing between grid points) tessellating the computational domain where the PDEs are being solved. For the purpose of this work, we assume an FEM-based discretization and the following generic error estimates.
Assumption 2.2.
Let π’ be the solution of the PDE equations in the forward problem. We assume that π’ β π , where π is a suitable vector space, e.g. a Sobolev space. The FEM approximation error is given by
β π’ β π’ β β π β€ πΆ β 2 β β ,
(10)
where β is the level of discretization (each level β halves the mesh size of the previous level β β 1 ) and the corresponding mesh size is β
2 β β .
Remark 2.1.
For simplicity, we did not include the error rate of time discretization. However, the time discretization error typically can be controlled by the discretization scheme to scale with the same rate of the spatial discretization. This will lead to the same convergence rate as in Assumption 2.2. Taking the finite-time two-dimensional Navier Stokes equation as an example, the error rate is β π’ β ( π‘ ) β π’ β β ( π‘ ) β π» 1 β€ πΆ β | β + Ξ β π‘ | with a β 1 -iso- β 2 / β 1 mixed Finite element discretization and Implicit/Explicit (IMEX) Euler time discretization scheme [MR4523340, MR2429876].
More details on error estimates for FEM can be found in [MR2050138, MR1043610], while for SEM and FVM, the interested reader may refer to [karniadakis2005spectral] and [MR2897628]. In analogy to Equation (7), we can define the numerical approximation of the forward map as follows
π’ β : π β β π β π³ β π ; π’ β ( π³ )
( β± 1 ( π’ β ( π³ ) , β± 2 ( π’ β ( π³ ) ) , β¦ , β± π ( π’ β ( π³ ) ) ) .
(11)
where π’ β is the solution of the discrete forward problem and β± π for π
1 , β¦ , π are π continuous bounded linear functionals. Thanks to Assumption 2.1, we can write the posterior probability measure also for the discrete problem we are considering here (in analogy with the continuous counterpart in Equation (8))
π β πΎ β , π¦ π β πΎ β exp β‘ ( β Ξ¦ β β ( π§ ; π¦ ) ) ,
(12)
where Ξ¦ β β ( π§ ; π¦ ) is the discrete potential function. Given Assumption 2.2 and Equation (9), it follows immediately that the Hellinger distance metric between the continuous posterior πΎ π¦ and the discrete one πΎ π¦ , β is bounded for every numerical refinement level β
π π» β ( πΎ π¦ , πΎ β , π¦ ) β€ πΆ β ( π¦ ) β 2 β β ,
(13)
where πΆ β ( π¦ ) is a positive constant, that depends only of the data π¦ .
Obviously, the larger the discretization level β (i.e. the finer the mesh), the more computationally expensive the problem. In fact, the number of degrees of freedom of the corresponding discrete linear system increases exponentially with respect to β . Hence, achieving a solution with a desired (and ideally small) error might be out of reach even with abundant computational resources. DL-based approximation methods (also referred to as DL-based surrogates) can come to the rescue here, and are introduced next.
2.2.2DL-based approximation methods
The second case considered uses DL models to accelerate the solution of Bayesian inverse problems by replacing the computationally expensive numerical approximation just introduced in Section 2.2.1 with their faster DL-based surrogate model counterparts. Let us denote π’ ~ : π β β π as a nonlinear map defined by a trained DL model. We assume that the DL model is trained with data generated with classical numerical methods, e.g. FEM, and that the objective is to solve the inverse problem with an error less than or equal to πͺ β ( 2 β πΏ ) . The procedure for solving such an inverse problem with DL-based surrogate acceleration is as follows.
First, we use a suitable numerical method (e.g. FEM, SEM, or FVM) to discretize the problem and generate the data. We assume that we use a characteristic mesh size equal to β
πͺ β ( 2 β πΏ ) ; to achieve the target accuracy, thanks to Assumption 2.2. Second, we use the generated numerical data as training data for the DL model. Third, we use the trained DL model as a surrogate to quickly run an MCMC chain. The estimated expectation of the QoI will be within the desired error if the DL-based surrogate model is as accurate as the numerical model.
However, empirically, the trained DL model can hardly achieve the same level of accuracy as the numerical approximation used to generate the training data, and will lead to additional error. We can formalize this statement as follows.
Assumption 2.3.
Given a DL model trained with data generated by a numerical approximation of the underlying forward problem that uses a mesh size β
2 β πΏ , and that has the error bound defined in Assumption 2.2, we can write
β πΊ ~ β ( π³ ) β π’ β π β€ πΆ β 2 β πΏ + π ,
(14)
where π³ is the input, and π accounts for the error of the DL model. In order for the DL error to be small, we require a small value of π .
In practice, we expect π to be small when we have a reasonably good DL model trained with sufficient data. However, in general, direct replacement of the numerical solver with a DL-based surrogate will lead to a posterior distribution estimation error of πͺ β ( 2 β πΏ + π ) given by Equation (13). Hence, producing an error gap to the targeted πͺ β ( 2 β πΏ ) .
In order to mitigate the shortcomings of DL-based surrogate models, one can refine the mesh of the numerical model used for data generation and increase the size of the training data to produce a possibly more accurate DL model that can reach the desired accuracy. However, that will increase the computational cost by many folds. In general, to solve a two-dimensional problem, the minimum increment of the computational cost of the numerical model is 4 times, and 8 times for three-dimensional problems, not to mention more challenging problems whose computational cost does not scale linearly with the degrees of freedom. In addition to the cost of finer numerical solvers, the larger amount of data will also increase the DL training cost. Even if we are willing to pay the cost, it was shown that there is an empirical limit to the accuracy that certain DL models can reach [CAO2023112104]. Therefore, in some cases, the desired accuracy may be unreachable by direct substitution of the numerical model with a DL surrogate. In the next section, we propose a different approach to correct the error statistically with the MCMC method.
2.3Hybrid two-level MCMC with uniform prior
In this section, we propose the hybrid two-level MCMC method for error correction of the DL-based surrogate model for Bayesian inverse problems. The new approach is inspired by the multilevel version of MCMC, which was shown to reduce the computational cost of standard MCMC for various problems by two orders of magnitude [MR3372290, MR3084684, MR4246090, MR4523340]. Our hybrid two-level method utilizes both the DL-based surrogate model and the high-fidelity numerical model to sample the posterior probability of Bayesian inverse problems. In particular, we run a base MCMC chain with a DL-based surrogate, and a short correction MCMC chain with a numerical model with known accuracy. The latter is deployed to correct the statistical error of the MCMC chain generated by the DL-based surrogate.
Numerical multi-level approaches have been very successful for multi-scale physical problems. However, implementing multi-level algorithm for generic engineering or scientific problems can be challenging due to complex meshes and instability of coarse numerical models. We see the potential to avoid those challenges by hybridizing the DL-based surrogate and numerical solvers under the same mathematical framework. Hence, we propose a two-level hybrid approach inspired by the telescoping argument of the multilevel Monte Carlo algorithm [MR2436856]. We note that another potential approach is to use the DL-based surrogate model as a filter for the MCMC sampler, as inspired by [MR2231730]. However, our proposed approach can be an alternative approach and potentially generalizable beyond the MCMC algorithm (i.e., our approach may also be applied to other methods such as the ensemble Kalman filter and sequential Monte Carlo methods, among others where a telescoping sum structure is applicable).
We now start introducing our hybrid two-level MCMC method. To this end, we denote the π β ( π§ ) as π for simplicity. We further indicate the posterior distribution approximated by the DL-based surrogate model as πΎ DL , and the numerically approximated posterior distribution as πΎ num . With the target precision of πͺ β ( 2 β πΏ ) and Assumption 2.3, πΎ num and πΎ DL are equivalent to πΎ πΏ , π¦ and πΎ πΏ β π , π¦ in Section 2.2. In our two-level approach, we can rewrite the numerical approximation of the expected QoI π as follows
πΌ πΎ num β [ π ]
πΌ πΎ num β [ π ] β πΌ πΎ DL β [ π ] + πΌ πΎ π· β πΏ β [ π ]
( πΌ πΎ num β πΌ πΎ DL ) β [ π ] + πΌ πΎ π· β πΏ β [ π ] .
(15)
To derive a computable estimator with MCMC chains, we observe that the first term on the right hand side in (2.3) can be transformed as follows
( πΌ πΎ num β πΌ πΎ DL ) β [ π ]
1 π num β β« π exp β‘ ( β Ξ¦ num ) β π β π πΎ β 1 π DL β β« π exp β‘ ( β Ξ¦ DL ) β π β π πΎ
1
π
num
β
β«
π
(
exp
β‘
(
β
Ξ¦
num
)
β
exp
β‘
(
β
Ξ¦
DL
)
)
β
π
β
π
πΎ
+
(
1
π
num
β
1
π
DL
)
β
β«
π
exp
β‘
(
β
Ξ¦
DL
)
β
π
β
π
πΎ
1 π num β β« π exp β‘ ( β Ξ¦ num ) β ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π β π πΎ
- ( π DL π num β 1 ) β 1 π DL β β« π exp β‘ ( β Ξ¦ DL ) β π β π πΎ ,
(16)
where π num
β« π exp β‘ ( β Ξ¦ num ) β π πΎ and π DL
β« π exp β‘ ( β Ξ¦ DL ) β π πΎ are the normalization constants. The constant ( π DL / π num β 1 ) can be expanded as
( π DL π num β 1 )
1 π num β β« π ( exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 ) β exp β‘ ( β Ξ¦ num ) β π πΎ .
(17)
We note that the integral 1 π num β β« π ( β ) β exp β‘ ( β Ξ¦ num ) β π πΎ and 1 π DL β β« π ( β ) β exp β‘ ( β Ξ¦ DL ) β π πΎ can be estimated with an MCMC estimator π π num πΎ num β [ β ] and π π DL πΎ DL β [ β ] , where the π num and π DL is the number of numerical MCMC samples and DL surrogate MCMC samples. Having defined equations (2.3), (2.3), and (17), we can write
πΌ πΎ num β [ π ]
πΌ πΎ num β [ ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π ]
- πΌ πΎ num β [ exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 ] β πΌ πΎ DL β [ π ]
- πΌ πΎ DL β [ π ] ,
(18)
and we can now define the hybrid two-level MCMC estimator π hybrid β [ π ] of πΌ πΎ π¦ β [ π ] as follows
π hybrid β [ π ]
π π num πΎ num β [ ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π ]
- π π num πΎ num β [ exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 ] β π π DL πΎ DL β [ π ]
- π π DL πΎ DL β [ π ] .
(19)
The new hybrid two-level MCMC approach for Bayesian inverse problems just introduced is simple, and it can therefore be adopted easily with legacy numerical models without too many changes in the code base. An important aspect of the new hybrid two-level MCMC introduced in equation (2.3) is its error analysis. In particular, we aim to show how the correction chain effectively corrects the estimator error caused by DL-based surrogate model.
Theorem 2.1.
The hybrid two-level MCMC estimator error can be decomposed into the following three components:
πΌ πΎ π¦ β [ π ] β π hybrid β [ π ]
I + II + III ,
(20a)
where
I := πΌ πΎ π¦ β [ π ] β πΌ πΎ num β [ π ]
(20b)
II := πΌ πΎ π· β πΏ β [ π ] β π π DL πΎ π· β πΏ β [ π ]
(20c)
III := πΌ πΎ num β [ ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π ]
β π π num πΎ num β [ ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π ]
- πΌ πΎ num β [ exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 ] β πΌ πΎ DL β [ π ]
β π π num πΎ num β [ exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 ] β π π DL πΎ DL β [ π ]
(20d) Proof.
Given the estimator error, we observe
πΌ πΎ π¦ β [ π ] β π hybrid β [ π ]
πΌ πΎ π¦ β [ π ] β πΌ πΎ num β [ π ] + πΌ πΎ num β [ π ] β π hybrid β [ π ]
I + πΌ πΎ num β [ π ] β π hybrid β [ π ]
(21)
With equation (2.3) and (2.3), we have
πΌ πΎ π¦ β [ π ] β π hybrid β [ π ]
I + πΌ πΎ num β [ π ] β π hybrid β [ π ]
I
+
πΌ
πΎ
π·
β
πΏ
β
[
π
]
+
πΌ
πΎ
num
β
[
(
1
β
exp
β‘
(
Ξ¦
num
β
Ξ¦
DL
)
)
β
π
]
+
πΌ
πΎ
num
β
[
exp
β‘
(
Ξ¦
num
β
Ξ¦
DL
)
β
1
]
β
πΌ
πΎ
DL
β
[
π
]
β
π
π
DL
πΎ
π·
β
πΏ
β
[
π
]
β
π
π
num
πΎ
num
β
[
(
1
β
exp
β‘
(
Ξ¦
num
β
Ξ¦
DL
)
)
β
π
]
β
π
π
num
πΎ
num
β
[
exp
β‘
(
Ξ¦
num
β
Ξ¦
DL
)
β
1
]
β
π
π
DL
πΎ
DL
β
[
π
]
I + II + III .
(22)
β
As shown, the overall error bound for our hybrid two-level MCMC method is composed of three error terms in equation (20). We analyse each error term individually, and assemble the overall error result as a conclusion to this analysis.
Proposition 2.1.
Let π
{ π³ ( π ) } π β β be a Markov chain, π« be the probability measure of the Markov chain. For every bounded π and every π β β , we have the following mean square error bound:
( β° β [ | πΌ β [ π β ( π§ ) ] β 1 π β β π
1 π π β ( π§ ( π ) ) | 2 ] ) 1 / 2 β€ πΆ β sup π§ β π | π β ( π§ ) | β π β 1 / 2 ,
where β° is the expectation over all realizations of π with respect to the measure π« .
This is a standard result from Markov chain theory, detailed proof can be found in [levin2017markov, MR3084684]. For the flow of the paper, we include the proposition here without proof.
Theorem 2.2.
We denote by π hybrid
{ π num , π DL } the collection of Markov chains obtained with numerical forward solver and DL-based surrogate solver Let π« num and π« DL be the probability measure of respective Markov chains, we denote π hybrid
π« num β β¨ π« DL . With π DL
πΆ DL β 2 2 β πΏ and π num
πΆ num β ( 1 + 2 π ) 2 , we have the following theoretical error estimate of our hybrid two-level MCMC approach under uniform priors
β° hybrid β [ | πΌ πΎ π¦ β [ π ] β π hybrid β [ π ] | ] β€ πΆ hybrid β 2 β πΏ ,
(23)
where β° hybrid is the expectation over all realizations of the collection π hybrid with respect to the product meansure π hybrid .
Proof.
From Theorem 2.1, we decompose the overall error into three components. For error term I, from equation (20) and equation (13), we can obtain the following error bound
| I | := | πΌ πΎ π¦ β [ π ] β πΌ πΎ num β [ π ] | β€ 2 β ( πΌ πΎ π¦ β [ π 2 ] + πΌ πΎ num β [ π 2 ] ) 1 / 2 β π π» β ( πΎ π¦ , πΎ num ) β€ πΆ β 2 β πΏ ,
(24)
where the details of the first inequality can be found in [MR2652785]. For error term II, from Proposition 2.1 we can obtain the following error bound
β° DL β [ | II | ] β€ ( β° DL β [ | II | 2 ] ) 1 / 2 := ( β° DL β [ | πΌ πΎ π· β πΏ β [ π ] β π π DL πΎ π· β πΏ β [ π ] | 2 ] ) 1 / 2 β€ πΆ β π π· β πΏ β 1 / 2 ,
(25)
where β° DL is the expectation over all realizations of Markov chain π DL with respect to the probability measure π« π· β πΏ .
Finally, for error term III, we can use the inequality | exp β‘ ( π₯ ) β exp β‘ ( π¦ ) | β€ | π₯ β π¦ | β ( exp β‘ ( π₯ ) + exp β‘ ( π¦ ) ) , to obtain
sup π§ β π | 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) |
β€ sup π§ β π | Ξ¦ num β Ξ¦ DL | β ( 1 + exp β‘ ( Ξ¦ num β Ξ¦ DL ) )
β€ πΆ β sup π§ β π ( | | π¦ β π’ num | 2 β | π¦ β π’ DL | 2 | )
β€ πΆ β sup π§ β π ( 2 β | π¦ | + | π’ num | + | π’ DL | ) β | π’ num β π’ π· β πΏ |
β€ πΆ β sup π§ β π ( | π’ num β π’ | + | π’ DL β π’ | )
β€ πΆ β ( 2 β πΏ + 2 β πΏ + π )
β€ πΆ β ( 1 + 2 π ) β 2 β πΏ ,
(26)
that leads to
β° num β [ { πΌ πΎ num β [ ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π ] β π π num πΎ num β [ ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π ] } 2 ]
β€ πΆ β π num β 1 β sup π§ β π ( | 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) | 2 )
β€ πΆ β π num β 1 β ( 1 + 2 π ) 2 β 2 β 2 β πΏ .
(27)
Similarly, we have
β° hybrid [ { πΌ πΎ num [ exp ( Ξ¦ num β Ξ¦ DL ) β 1 ] β πΌ πΎ DL [ π ]
β π π num πΎ num [ exp ( Ξ¦ num β Ξ¦ DL ) β 1 ] β π π DL πΎ DL [ π ] } 2 ]
β€
πΆ β° num [ { πΌ πΎ num [ exp ( Ξ¦ num β Ξ¦ DL ) β 1 ]
β π π num πΎ num [ exp ( Ξ¦ num β Ξ¦ DL ) β 1 ] } 2 ] β sup π§ β π | π | 2
- πΆ β sup π§ β π | exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 | 2 β β° DL β [ { πΌ πΎ DL β [ π ] β π π DL πΎ DL β [ π ] } 2 ]
β€
πΆ β π num β 1 β sup π§ β π | exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 | 2 + πΆ β π DL β 1 β sup π§ β π | exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 | 2 .
β€
πΆ β π num β 1 β ( 1 + 2 π ) 2 β 2 β 2 β πΏ + πΆ β π DL β 1 β ( 1 + 2 π ) 2 β 2 β 2 β πΏ .
(28)
By combining equations (2.3) and (2.3), we have the overall error estimate for III, that is
β° hybrid β | III | 2
:= β° hybrid [ | πΌ πΎ num [ ( 1 β exp ( Ξ¦ num β Ξ¦ DL ) ) π ]
β π π num πΎ num β [ ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π ]
- πΌ πΎ num β [ exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 ] β πΌ πΎ DL β [ π ]
β π π num πΎ num [ exp ( Ξ¦ num β Ξ¦ DL ) β 1 ] β π π DL πΎ DL [ π ] | 2 ]
β€ πΆ β π num β 1 β ( 1 + 2 π ) 2 β 2 β 2 β πΏ + π DL β 1 β ( 1 + 2 π ) 2 β 2 β 2 β πΏ .
(29)
Until now, we are still free to choose the number of samples for π num and π DL . To balance errors I, II and III, we can choose the sampling number π DL
πΆ π· β πΏ β 2 2 β πΏ (see Equation 25) and π num
πΆ num β ( 1 + 2 π ) 2 , that allows us to write the overall error estimate. Here, we differentiate the two constants of different values by πΆ num and πΆ DL . Finally we have a final numerical to DL sample ratio of πΆ num πΆ DL β ( 1 + 2 π ) 2 / 2 2 β πΏ . β
With the DL-based surrogate sample number π DL
πΆ DL β 2 2 β πΏ and numerical sample number π num
πΆ num β ( 1 + 2 π ) , where both πΆ does not depend on πΏ and π , Theorem 2.2 shows that there is an optimal ratio πΆ num β ( 1 + 2 π ) 2 / πΆ DL β 2 2 β πΏ for our hybrid method to reach the same accuracy as plain MCMC with a numerical model. However it is theoretically non-trivial to work out both constantsβ values. For high fidelity problems which benefit the most from the adoption of a DL-based surrogate model, a large L value is expected in which the ( 1 + 2 π ) 2 / 2 2 β πΏ term will be the dominant factor leading to a small overall ratio. Nevertheless, the theory is only good enough to give a general guideline. Exact optimal ratio is not available analytically with dependencies on factors other than πΏ and π , such as regularity of specific problems. We propose a sweep test for our numerical experiments and potential practical applications, where we empirically test a range of different ratios to get the best hybrid configuration. In addition, if we assume the computational speedup rate of the DL-based surrogate model as π
π‘ num / π‘ DL , where π‘ num is the average computational time for one forward solve with numerical model and π‘ DL is the computational time of one DL-based surrogate evaluation time, we can estimate the overall speedup of our new hybrid two-level approach compared to the plain numerical MCMC. In particular, with the choice of samples π DL and π num , the overall speedup is πͺ β ( 2 2 β πΏ / ( 1 π β 2 2 β πΏ + πΆ num πΆ DL β ( 1 + 2 π ) 2 ) ) compared to the plain numerical MCMC. We note that the hybrid two-level approach can be easily parallelized with two processes, thus the actual speedup is πͺ β ( 2 2 β πΏ / max β‘ ( 1 π β 2 2 β πΏ , πΆ num πΆ DL β ( 1 + 2 π ) 2 ) ) . Even though the discussion is based on the setup of a uniform prior, the same conclusions hold for Gaussian priors. Details of the latter are provided in A. In Section 3, we show several experiments that validate the theoretical findings.
As a final note, we shall point out that, even though the hybrid method is able to provide accurate posterior expectations of the quantities of interest and the variance (which can be computed from the expectations), the method will not generate a large number of actual numerical samples from the highly accurate numerically approximated posterior distribution (the exact reason why computational cost is saved). This limits the method from producing a histogram like a conventional MCMC chain. However, a less accurate histogram can still be generated from the DL surrogate samples.
3Numerical experiments
After having introduced the new hybrid two-level MCMC approach, we now present some numerical experiments to validate the theoretical error estimates and to understand the computational performance of the new approach. The numerical experiments span three different PDE problems, namely a Poisson equation, a nonlinear reaction-diffusion equation, and a Navier-Stokes equation. These three problems include both elliptic and parabolic differential problems, with different levels of complexity, and constitute an established benchmark for testing novel methods in the context of Bayesian inverse problems [VillaPetraGhattas21].
In the numerical experiments, we consider both uniform and Gaussian priors, to demonstrate the theoretical error estimates presented in Sections 2.3, and A, respectively.
(a)Mesh A: 32 Γ 32 resolution (b)Mesh B: 64 Γ 64 resolution Figure 2:Meshes adopted and points where the solution is observed (red crosses). Mesh A is used for the numerical experiments on a Poisson and a nonlinear reaction-diffusion equations, while Mesh B is used for the numerical experiments on a Navier-Stokes equation.
All results are obtained in the two-dimensional squared computational domains (or meshes) π· β β [ 0 , 1 ] Γ [ 0 , 1 ] , Mesh A and Mesh B, depicted in Figure 2, where we observe the solution π’ in equally spaced fixed positions, highlighted as red crosses. Mesh A in Figure 2a is used for the numerical experiments with a Poisson equation and a nonlinear reaction-diffusion equation, and has a number of cells equal to 32 Γ 32
1024 (mesh level β
5 ). Mesh B in Figure 2b is used for the numerical experiment with a Navier-Stokes equation, and has a number of cells 64 Γ 64
4096 (mesh level β
6 ).
For the different problems considered, we used a range of different neural network architectures, to show that the method proposed here is agnostic to the choice of DL surrogate. The list of experiments carried out in the following subsections is summarized in Table 1.
1 2 3 4 5 6
Problem Poisson Equation Reaction Diffusion Navier Stokes Prior Uniform Gaussian Uniform Gaussian Uniform Gaussian DL Model FCN CNN GNN U-Net DeepONet FNO Section 3.1.1 3.1.2 3.2.1 3.2.1 3.3.1 3.3.1 Table 1:List of numerical experiments
We use Metropolis-Hasting algorithm with prior distribution as our proposal density for all our MCMC chains in the numerical experiments.
3.1Poisson equation
We first consider a Bayesian inverse problem with a forward model governed by the Poisson equation in the two-dimensional computational domain depicted in Figure 2a.
{ β β ( πΎ β ( π§ ) β β π’ β ( π₯ ) )
cos β‘ ( 2 β π β π₯ 1 ) β sin β‘ ( 2 β π β π₯ 2 ) ,
π’ β ( π₯ 1
0 )
0 ,
π’ β ( π₯ 1
1 )
1 ,
β π’ β π₯ 2 β ( π₯ 2
0 )
β π’ β π₯ 2 β ( π₯ 2
1 )
0 .
(30)
The data π’ is observed at thirty-six equally-distanced positions, as shown in figure 2a, using a random realization of the forward model with additive Gaussian noise πΏ that has zero mean and variance π 2
0.001 . We next present the uniform and Gaussian prior cases.
3.1.1Uniform prior
For this first case, we set the QoI to be the random field πΎ
π§ β cos β‘ ( 2 β π β π₯ 1 ) β sin β‘ ( 2 β π β π₯ 2 ) + 2.0 , whereby its prior distribution is uniform, namely π§ βΌ π β [ 0 , 1 ] . We solve equation (30) using FEM on Mesh A (depicted in Figure 2a), and randomly generate 8000 solution samples. The 8000 samples are partitioned into 4000 training samples, 2000 validation samples, and 2000 test samples to train a fully connected ReLU neural network. For details on the FEM solver and the neural network model, the interested reader may refer to B.1.
We performed numerical experiments on three setups: (i) a plain MCMC chain with numerical solver, denoted as Numerical MCMC in Table 2, (ii) a plain MCMC chain with DL-based surrogate model, denoted as DL MCMC in Table 2, (iii) and the proposed hybrid approach, denoted as Hybrid MCMC in Table 2, where we tested different lengths of the numerical samples chain, ranging from 1% to 100% of the total number of DL samples (see Fig 3). We show in detail two cases in Table 2, one with 1 % numerical samples and the other with 5 % numerical samples (compared to the total number of DL samples). All values showed in Table 2 are average results of eight MCMC runs.
Method Numerical MCMC DL MCMC Hybrid MCMC Hybrid MCMC Samples 100,000 100,000 100,000 +1,000 100,000 +5,000 Estimator error 9.72E-4 7.524E-3 8.75E-4 8.69E-4 Compute time [s] 1090.95 65.77 66.68 (serial) 65.77 (parallel) 120.32 (serial) 65.77 (parallel) Speed up - 16.59x 16.36 (serial) 16.59x (parallel) 9.07x (serial) 16.59x (parallel) Table 2:Estimator error and compute speedup for the elliptic problem with uniform prior
The reference value computed employing the FEM model and approximating the expected value of the posterior with a quadrature scheme with 32 Gaussian Legendre quadrature over the parametric domain π§ β [ 0 , 1 ] is also included. Quadrature estimation of the posterior mean shows a value of 0.313237 with mesh level β
10 β ( 1024 Γ 1024 β cells ) . This provides a highly accurate posterior expectation which can be considered as a true reference. Hence, all the estimator errors are calculated by comparing to this reference. We observe how the proposed hybrid two-level MCMC approach provides results that are comparable to the numerical MCMC method. The comparison between the two hybrid MCMC experiments also qualitatively validates Theorem 2.2, where an optimal ratio exists and additional numerical samples do not further improve the results.
Figure 3:Estimator error with different percentage of numerical sample against DL-based surrogate samples for the elliptic experiment with uniform prior
In addition to the results presented in Table 2, we also experimentally estimate the mean of the surrogate model error and numerical model error with respect to z again by a Gaussian Legendre quadrature (this time with 64 quadrature points), that in turn allows us to approximate π in Assumption 2.2. Using again the mesh at level β
10 as reference, we obtain π β β π’ πΏ
10 β π’ πΏ
5 β πΏ β β ( π· ) β 2.0 β πΈ β 4 , and of π β β π’ πΏ
10 β π’ πΏ
5 β πΏ β β ( π· ) β 1.0 β πΈ β 3 for the numerical forward solver and for the DL surrogate, respectively. With such, we have a rough estimate of π
2.3 . This is not a rigorous error bound estimate; nevertheless, it still provides a useful indication of the accuracy of the DL-based surrogate model. Follow Theorem 2.2, our estimate provides the ratio ( 1 + 2 π ) 2 / 2 2 β πΏ . In this specific case we obtain a ratio of approximately 4 % . It means that we require 4% numerical simulations in the correction chain to get an error with approximately the same order of magnitude of a full numerical MCMC chain, if πΆ num β πΆ DL . In practice, obtaining π may be challenging and constants πΆ num and πΆ DL are generally unknown, it may be more advantageous to compute π num / π DL empirically. Indeed, this ratio is application and user dependent. For instance, if we can afford more numerical forward simulations, we could potentially run more of them, albeit at higher computational costs. If we instead have a limited number of computational resources, and this is our primary constraint, we may want to limit the number of numerical forward simulations, while potentially accepting a higher error. This trade-off is depicted in Fig. 3, where we show the error with respect to the percentage of numerical samples against DL-based surrogate samples. We observe that the error reduces significantly with small number of numerical samples. However the error reduction by further increasing the ratio of numerical samples, even to 100%, is not as significant.
In our experiments, we ran very long (100,000 samples) base MCMC chains with DL-based surrogate models to ensure convergence. For completeness, we include additional diagnostic details of the MCMC experiments in C. First we show the trace plots and histogram plots of the MCMC chains with DL-based surrogate models in Fig 13 and Fig 14. The trace plots shows great mixing of samples from the posterior distribution. The histograms from different experiments show similar posterior distribution. We also show a between chains comparison of sample mean with the eight independent experiments starting from random initial sample in Fig 15a. The plots show clear convergence of the sample mean. We also calculated the Effective Sample Size (ESS). The eight experiments show an average ESS of 30,842 out of 100,000 MCMC samples. The sample mean convergence between eight MCMC chain and the Potential Scale Reduction Factor (PSRF) is shown in Fig 15b. The final PSRF, also known as R-hat, from Gelman-Rubin diagnostic is 1.00003. A value smaller than 1.2 is often considered as a good indication of convergence [MR1665662, gelman1992inference]. We also show the trace plot, autocorrelation function and histogram of the QoI in the correction chain in Fig 16, Fig 17, Fig 18 and Fig 19, where we observe a smaller variance compared with the base MCMC samples. Sample mean and PSRF plots are shown in Fig 20 and Fig 21, where the PSRF value for both QoI are 1.00209 and 1.00238. All the diagnostics show good indication of convergence.
Finally, we estimated the overall computational time of the hybrid two-level MCMC approach compared to the numerical MCMC and DL MCMC. The estimation is based on the average runtime for one numerical sample and one DL-based surrogate sample. The FEM solver uses Intel Xeon E5-2620 CPU, while the DL-based surrogate model evaluations used one NVIDIA RTX A6000 GPU (we use the same CPU and GPU specs for all the subsequent experiments). Due to the fact that the two MCMC chains in the hybrid MCMC method can be run concurrently, the hybrid MCMC with 100,000 surrogate samples and 5,000 numerical samples achieved the same speed-up as the plain MCMC with purely surrogate samples, but achieved a smaller error.
3.1.2Gaussian (log-normal) prior
In contrast to the uniform prior case just shown, for this experiment, the QoI πΎ is spatially varying, that is: πΎ
πΎ β ( π₯ ) , and we assume its prior distribution to be sampled from the following bi-Laplacian Gaussian prior
π β π
{ πΎ β β β ( Ξ β β π ) + πΏ β π
in β π·
( Ξ β β π ) β π + π½ β π
on β β π· ,
(31)
where π is the normal direction with respect to the boundary, πΎ
0.1 , πΏ
0.5 , π½
πΎ β πΏ and Ξ controls the anisotropicity where we use an identity matrix. In the numerical experiments, we sample π to obtain the non-negative random field sample πΎ β ( π₯ )
exp β‘ ( π ) by solving the equation π β π
π with FEM method on mesh 2a, where π is sampled from white noise. This is equivalent to sample from the bi-Laplacian covariance operator π β 2 . In this numerical experiment, the discretized πΎ β ( π₯ ) with finite dimension 32 Γ 32
1024 can be considered as the π§ in π β ( π§ ) in the preceding discussions. For more details of this particular prior setup, one can refer to Section 5.1.1 in [VillaPetraGhattas21] and references therein. We show four examples of random samples generated in this experiment in Figure 4.
(a)Sample 1. (b)Sample 2. (c)Sample 3. (d)Sample 4. Figure 4:Samples obtained from the bi-Laplacian random field in equation (31) for the Poisson equation with Gaussian prior.
Similarly to the uniform-prior case, we solve the forward problem in equation (30) via FEM on Mesh A (Figure 2a), and generate 8000 random samples. Again, the 8000 samples are partitioned into 4000 training samples, 2000 validation samples, and 2000 test samples to train a convolutional neural network. The convolutional neural network consists of 3 encoding layer, 1 fully connected layer and 3 decoding layers, and it is trained using the Adam optimizer for 10000 epochs. We have a rough estimate of π
0.228 . With reference to Theorem 2.2 assuming πΆ num β πΆ DL , the correction chain needs around 1% numerical samples compared to the long DL-based MCMC chain. Again, this is not a rigorous estimate of the DL model error, where the ratio is an underestimation. For a more accurate estimation of percentage of numerical samples needed, we include Fig 5 to show the error with respect to the percentage of numerical samples against DL-based surrogate samples. Fig 5 shows the error reduced significantly with 5% or more numerical samples. We remark that the error used here is the πΏ 2 and πΏ β difference with respect to the classical numerical MCMC results, which is different from the error computed against a highly accurate quadrature in Fig 3. This is due to the computational challenge to compute quadrature of a high dimensional problem like this experiment. In the subsequent numerical experiment, we stick to the same approach.
In analogy with the uniform-prior case, we run three experiments: (i) a plain MCMC chain with numerical solver, denoted as Numerical MCMC in Table 3, (ii) a plain MCMC chain with DL-based surrogate model, denoted as DL MCMC in Table 3, and the proposed hybrid approach, denoted as Hybrid MCMC in Table 3. We show in detail three cases in Table 3, namely π num / π DL
1 % , 5 % , 10 % .
Figure 5:Error in comparison with numerical MCMC results at different percentage of numerical samples against DL-based surrogate samples for the elliptic experiment with Gaussian prior
Due to the high dimensional nature of the random field samples, we will not include all the trace plots and histogram of each pixel to show the diagnostics of each MCMC chain. We simply report that the max PSRF of the DL-based surrogate MCMC chain of 100,000 samples converged, being 1.00000714 < 1.2 . The max PSRF for the QoI of A1, A2, β¦, A8 (refer to the hybrid method for Gaussian prior in A) in the correction chain of 1000 samples are 1.00794, 1.00744, 1.00558, 1.00413, 1.00231, 1.01205, 1.00962 and 1.00007. All PSRF values for the correction MCMC chain show good indication of convergence.
The average results of eight MCMC runs are depicted in Figure 6. The expectation of the posterior from the MCMC chains generated solely with a DL-based surrogate model has obvious discrepancies with the plain MCMC chains generated solely with a numerical solver (that constitute the reference). Our hybrid two-level MCMC approach, with the addition of only few numerical samples, is able to significantly improve the results making it comparable to the reference, at a fraction of the computational cost.
We summarize the results in Table 3, where the πΏ 2 and πΏ β difference between the DL-based surrogate accelerated MCMC results and classical numerical MCMC results are presented. From the results presented, there are significant improvements in terms of accuracy with the hybrid approach. With only 1 % additional numerical samples on top of the DL-based MCMC chain, the results get much closer to the one from the numerical MCMC. With 5% and 10% additional numerical samples, the results of hybrid approach get even closer to the single chain numerical MCMC result.
(a)Numerical-only (100000). (b)DL-only (100000). (c)Hybrid (100000+1000). (d)Hybrid (100000+5000). (e)Hybrid (100000+10000). Figure 6:Expected mean of πΎ β ( π₯ ) from eight runs of MCMC with elliptic equation with Gaussian prior. Method Numerical MCMC DL MCMC Hybrid MCMC Hybrid MCMC Hybrid MCMC Samples 100,000 100,000 100,000
- 1,000 100,000
- 5,000 100,000
- 10,000
πΏ 2 difference - 9.374E-3 2.997E-3 1.562E-3 1.114E-3
πΏ β difference - 4.834E-1 2.944E-1 1.244E-1 9.407E-2 Compute time [s] 1080.67 70.54 81.35 (serial) 70.54 (parallel) 124.57 (serial) 70.54 (parallel) 178.61 (serial) 108.07 (parallel) Speed up - 15.32x 13.28x (serial) 15.32x (parallel) 8.67x (serial) 15.32x (parallel) 6.05x (serial) 10x (parallel) Table 3:Difference between the posterior expectation results of plain numerical MCMC and the posterior expectation results of the plain and hybrid DL-based MCMC with elliptic equation with Gaussian prior 3.2Nonlinear reaction-diffusion equation
We consider the Bayesian inverse problem with the forward model governed by the following nonlinear reaction-diffusion equation in a two-dimensional unit square domain π·
{ β β ( πΎ β ( π₯ ) β β π’ β ( π₯ ) ) + π’ 3
0 ,
π’ β ( π₯ 1
0 )
0 ,
π’ β ( π₯ 1
1 )
1 ,
β π’ β π₯ 2 β ( π₯ 2
0 )
β π’ β π₯ 2 β ( π₯ 2
1 )
0 .
(32)
Thirty-six equally distanced observations are captured from a random realization of the forward model with additional Gaussian noise πΏ with zero mean and variance π 2
0.1 .
3.2.1Uniform prior
In this section, we consider the uniform prior case with a random field πΎ β ( π₯ ) that depends on uniformly distributed coefficients π§ π , π
0 , 1 , β¦ , 4
ln β‘ ( πΎ β ( π₯ ) )
π§ 0
π§ 1 cos ( 2 π π₯ 1 ) sin ( 2 π π₯ 2 )
π§ 2 sin ( 2 π π₯ 1 ) cos ( 2 π π₯ 2 ) )
π§ 3 cos ( 2 π π₯ 1 ) cos ( 2 π π₯ 2 )
π§ 4 sin ( 2 π π₯ 1 ) sin ( 2 π π₯ 2 ) ) ,
where π§ π βΌ π β [ β 1 , 1 ] , π
0 , 1 , β¦ , 4 . We solve the reaction diffusion equation (32) with the FEM method on a 32 Γ 32 uniformly spaced Mesh A 2a. We randomly generated 4000 samples with a FEM numerical solver. The 4000 data are partitioned into 2000 training samples, 1000 validation samples, and 1000 test samples. We train a vanilla Message Passing Graph Neural Network (MPGNN) as our DL surrogate. The details of the FEM solver and the Graph Neural Network architecture can be found in B.2. We have a rough estimate of π
3.91 . With reference to Theorem 2.2 assuming πΆ num β πΆ DL , the correction chain only needs around 26% numerical samples compared to the long DL-based MCMC chain. Yet again this is not a rigorous error rate for the DL model, where the ratio is an overestimation. For a more accurate estimation of numerical samples needed, we include Fig 7 to show the error with respect to the percentage of numerical samples against DL-based surrogate samples. Fig 7 shows the error reduced significantly even with a small number of numerical samples.
Figure 7:Error in comparison with numerical MCMC results at different percentage of numerical samples against DL-based surrogate samples
We run three experiments: (i) a plain MCMC chain with numerical solver, denoted as Numerical MCMC in Table 4, (ii) a plain MCMC chain with the MPGNN-based DL surrogate model, denoted as DL MCMC in Table 4, and the proposed hybrid approach, denoted as Hybrid MCMC in Table 4, with the number of samples in the numerical chain being a fraction of the samples in the DL chain. We tested π num / π DL ratios ranging from 1% to 100% as shown in Fig 7. We show in details three cases ( 1 % , 5%, and 10%) in Fig 8 and Table 4. For compactness of the paper we skip trace plots and histogram plots. We computed PSRF for all QoIs in both MCMC chains. The results show maximum PSRF value of 1.00922 which is smaller than the common indicative 1.2 value.
The average results of five MCMC runs are presented in Figure 8. The πΏ 2 and πΏ β difference between the numerical MCMC results and the DL-based methods (including the DL MCMC and the Hybrid MCMC) are presented in Table 4. These show that the additional numerical samples in the hybrid method reduce both πΏ 2 and πΏ β difference compared to the DL MCMC method. The results validate the theoretical conclusion in Theorem 2.2, that the hybrid two-level approach can reach the same accuracy level as the numerical MCMC at a fraction of the computational cost.
(a)Numerical-only (100000) (b)DL-only (100000) (c)Hybrid (100000+1000) (d)Hybrid (100000+5000) (e)Hybrid (100000+10000) Figure 8:Expected mean of πΎ β ( π₯ ) from eight runs of MCMC with reaction diffusion equation with uniform prior. Method Numerical MCMC DL MCMC Hybrid MCMC Hybrid MCMC Hybrid MCMC Samples 100,000 100,000 100,000
- 1,000 100,000
- 5,000 100,000
- 10,000
πΏ 2 difference - 2.914E-3 9.638E-4 4.756E-4 3.861E-4
πΏ β difference - 1.772E-1 0.716E-2 0.351E-2 0.285E-2 Compute time [s] 2840.75 65.78 94.19 (serial) 65.78 (parallel) 207.82 (serial) 142.04 (parallel) 349.86 (serial) 284.08 (parallel) Speed up - 43.19x 30.15x (serial) 43.19x (parallel) 13.67x (serial) 20x (parallel) 8.12x (serial) 10x (parallel) Table 4:Difference between the posterior expectation results of plain numerical MCMC and the posterior expectation results of the plain and hybrid DL-based MCMC with non-linear reaction-diffusion equation with uniform prior 3.2.2Gaussian prior
We follow the same random field setup for πΎ β ( π₯ ) in Section 3.1.2. We randomly generated 4000 samples with the Finite Element solver. The 4000 data are partitioned into 2000 training data, 1000 validation data, and 1000 test data to train a U-net neural network. The details of the FEM solver and the neural network architecture can be found in B.2. We have a rough estimate of π
0.93 . With reference to Theorem 2.2 assuming πΆ num β πΆ DL , the correction chain only needs around 1% numerical samples compared to the long DL-based MCMC chain. For a more accurate estimation, we include Fig 9 to show the error with respect to the percentage of numerical samples. We see that the estimation π we have is not perfect, where in the sweep test we see an increase of the error with 1% of numerical samples but better results are acheived with 5% or more numerical samples.
Figure 9:Error in comparison with numerical MCMC results at different percentage of numerical samples against DL-based surrogate samples for the reaction diffusion experiments with Gaussian prior
We performed numerical experiments on three setups: (i) a plain MCMC chain with numerical solver, denoted as Numerical MCMC in Table 5, (ii) a plain MCMC chain with DL-based surrogate model, denoted as DL MCMC in Table 5, (iii) and the proposed hybrid approach, denoted as Hybrid MCMC in Table 5, once again testing cases such that the π num / π DL ratio ranges from 1% to 100%, as shown in Fig 9. We show in detail three cases (1%, 5%, and 10%) in Table 5. For compactness of the paper we skip trace plots and histogram plots. We computed PSRF for all QoIs in both MCMC chains. The results show maximum PSRF value of 1.0131 which is smaller than the common indicative 1.2 value.
The results of the average of the eight MCMC run are shown in Figure 10. The πΏ 2 and πΏ β difference between MCMC results generated with the DL-based surrogate and plain numerical MCMC are included in Table 5. The results show that the additional numerical samples in the hybrid method reduce both πΏ 2 and πΏ β difference compared to the plain MCMC model.
(a)Numerical MCMC(100,000) (b)DL MCMC (100,000) (c)Hybrid (100,000+1,000) (d)Hybrid (100,000+5,000) (e)Hybrid (100,000+10,000) Figure 10:Expected mean of πΎ β ( π₯ ) from eight runs of MCMC for the nonlinear reaction-diffusion experiment with Gaussian prior Method Numerical MCMC DL MCMC Hybrid MCMC Hybrid MCMC Hybrid MCMC Samples 100,000 100,000 100,000
- 1,000 100,000
- 5,000 100,000
- 10,000
πΏ 2 difference - 5.953E-3 6.466E-3 2.643E-3 1.565E-3
πΏ β difference - 3.203E-1 7.151E-1 2.296E-2 1.448E-2 Compute time [s] 3396.72 178.81 212.78 (serial) 178.81 (parallel) 348.65 (serial) 178.81 (parallel) 518.78 (serial) 339.67 (parallel) Speed up - 18.99x 15.96x (serial) 18.99x (parallel) 9.74x (serial) 18.99x (parallel) 6.55x (serial) 10.0x (parallel) Table 5:Difference between the posterior expectation results of plain numerical MCMC and the posterior expectation results of the plain and hybrid DL-based MCMC with non-linear reaction-diffusion equation with Gaussian prior 3.3Navier Stokes equations
We consider the Bayesian inverse problem with the forward model governed by the two dimensional Navier Stokes equations in the vorticity form in a domain of two-dimensional unit torus π 2 ,
{ β π β ( π₯ , π‘ ) β π‘ + π’ β ( π₯ , π‘ ) β β π β ( π₯ , π‘ ) β π β Ξ β π β ( π₯ , π‘ )
π ,
for β π₯ β π 2 ,
β β π’ β ( π₯ , π‘ )
0 ,
for β π₯ β π 2 ,
π β ( π₯ , 0 )
π 0 ;
(33)
with periodic boundary conditions and forcing
π
0.1 β ( sin β‘ ( 2 β π β ( π₯ 1 + π₯ 2 ) ) + cos β‘ ( 2 β π β ( π₯ 1 + π₯ 2 ) ) ) .
(34)
π is the vorticity, π’ is the velocity, and π
0.001 is the viscosity. Thirty-six equally distanced observations are captured from a random realization of the forward model with additional Gaussian noise πΏ with zero mean and variance π 2
1 .
3.3.1Uniform prior
In this section, we consider the uniform prior case, using a Fourier expansion of π 0 with coefficient of each Fourier term uniformly distributed
π π β π
π§ π β π β π 2 β 2 β 7 3 / 2 β ( 4 β π 2 β ( π 2 + π 2 ) + 49 ) β 5 / 4 ,
where π
64 is the number of modes in each axis, m and n are the mode index in the π₯ and π¦ directions, and π§ π β π βΌ π β [ β 1 , 1 ] . Using the inverse fast Fourier transform (ifft), we have
π 0 β ( π₯ , π¦ )
ifft β ( Z )
1 N 2 β β m
0 N β 1 β n
0 N β 1 Z mn β exp β‘ ( i2 β π β mx / N + i2 β π β ny / N ) .
This particular setup is chosen to match the Gaussian prior π© β ( 0 , 7 3 2 β ( β Ξ + 49 β πΌ ) β 2.5 ) used later in the Gaussian case. In the Gaussian prior case, The coefficient π§ π β π follows π β ( 0 , 1 ) instead of an uniform prior. We solve the two-dimensional Navier-Stokes equations (33) by using a pseudospectral method with Crank-Nicolson time integration on Mesh B, that is composed of 64 Γ 64 collocation points. We randomly generated 8000 samples with the numerical solver, where 4000 samples are used for training, 2000 for validation, and 2000 for testing. DeepONet is used as the DL-based surrogate model. For more details on the numerical solver and the DL architecture, the interested reader can refer to B.3. We have a rough estimate of π
1.93 . With reference to Theorem 2.2 assuming πΆ num β πΆ DL , the correction chain needs less than 1% numerical samples compared to the long DL-based MCMC chain. We also include Fig 11 to show the error with respect to the percentage of numerical samples. In the sweep test, we see an increase of the error with 1% of numerical samples but slightly better results are achieved with 5% or more numerical samples.
Figure 11:Error in comparison with numerical MCMC results at different percentage of numerical samples against DL-based surrogate samples for the 2D Navier Stokes experiments with uniform prior
We run three experiments: (i) a plain MCMC chain with numerical solver, denoted as Numerical MCMC in Table 6, (ii) a plain MCMC chain with DL-based surrogate model, denoted as DL MCMC in Table 6, and the proposed hybrid approach, denoted as Hybrid MCMC in Table 6. As usual, we tested rations π num / π DL from 1% to 100%, as shown in Fig. 11. We show in details three cases (1%, 5%, and 10%) in Table 6. For compactness of the paper we skip trace plots and histogram plots. We computed PSRF for all QoIs in both MCMC chains. The results show maximum PSRF value of 1.029 which is smaller than the common indicative 1.2 value. The average results of eight MCMC runs are included in Table 6. We note that in this case the DL MCMC based on DeepONet already achieves relatively good performance, and it is obviously the fastest method. Only marginal accuracy gains are achieved by using our hybrid approach with more than 5% of numerical samples. This can be the case when the DL-surrogate model approximates the numerical forward operator well, such that adding some numerical samples within our hybrid MCMC framework does not significantly improve the results.
Method Numerical MCMC DL MCMC Hybrid MCMC Hybrid MCMC Hybrid MCMC Samples 100,000 100,000 100,000
- 1,000 100,000
- 5,000 100,000
- 10,000
πΏ 2 difference - 1.453E-05 1.249E-05 1.171E-05 1.125E-5
πΏ β difference - 2.594E-3 2.278E-3 2.2.061E-3 1.976E-3 Compute time [s] 79490.13 846.78 1641.68 (serial) 846.78 (parallel) 4821.29 (serial) 3974.51 (parallel) 8795.79 (serial) 7949.01 (parallel) Speed up - 93.87x 48.42x (serial) 93.87x (parallel) 16.48x (serial) 20x (parallel) 9.37x (serial) 10x (parallel) Table 6:Difference between the posterior expectation results of plain numerical MCMC and the posterior expectation results of the plain and hybrid DL-based MCMC with Navier Stokes equations with uniform prior 3.3.2Gaussian prior
In this section, we consider the Gaussian prior case, focusing on π 0 . We sample the Gaussian random field from the following Gaussian prior with the distribution: π© β ( 0 , 7 3 2 β ( β Ξ + 49 β πΌ ) β 2.5 ) . We solve the above equation with the pseudo-spectral method with the Crank-Nicolson time integration method. In this experiment a resolution of 64 Γ 64 is used. We randomly generated 4000 samples with the numerical solver. The 4000 data are partitioned into 2000 training data, 1000 validation data, and 1000 test data to train the Fourier Neural Operator (FNO) as described in [DBLP:conf/iclr/LiKALBSA21]. Details of the DL-based surrogate model setup can be found in B.3. We have a rough estimate of π
3.65 . With reference to Theorem 2.2 assuming πΆ num β πΆ DL , the correction chain needs only around 5% numerical samples if compared to the long DL-based MCMC chain.
Figure 12:Error in comparison with numerical MCMC results at different percentage of numerical samples against DL-based surrogate samples for the 2D Navier Stokes experiments with Gaussian prior
We run three experiments: (i) a plain MCMC chain with numerical solver, denoted as Numerical MCMC in Table 7, (ii) a plain MCMC chain with DL-based surrogate model, denoted as DL MCMC in Table 7, and the proposed hybrid approach, denoted as Hybrid MCMC in Table 7, where we tested cases with the ratio π num / π DL ranging from 1% to 100%, as shown in Fig 12. We show in details three cases (1%, 5%, and 10%) in Table 7. For compactness of the paper we skip trace plots and histogram plots. We computed PSRF for all QoIs in both MCMC chains. The results show maximum PSRF value of 1.089 which is smaller than the common indicative 1.2 value.
The results of the average of the eight MCMC run are presented in Table 7 where the posterior expectation obtained from the hybrid MCMC algorithm is closer to the posterior expectation obtained from plain numerical MCMC, but with much lower computational cost.
Method Numerical MCMC DL MCMC Hybrid MCMC Hybrid MCMC Hybrid MCMC Samples 100,000 100,000 100,000
- 1,000 100,000
- 5,000 100,000
- 10,000
πΏ 2 difference N.A. 2.337E-4 1.934E-4 1.004E-4 6.995-5
πΏ β difference N.A. 2.869E-2 3.548E-2 1.772E-2 1.205E-2 Compute time [s] 92343.7 456.4 1379.8 (serial) 923.4 (parallel) 5082.6 (serial) 4617.2 (parallel) 9690.8 (serial) 9234.4 (parallel) Speed up - 202.29x 66.96x (serial) 100x (parallel) 18.17x (serial) 20x (parallel) 9.53x (serial) 10x (parallel) Table 7:Difference between the posterior expectation results of plain numerical MCMC and the posterior expectation results of the plain and hybrid DL-based MCMC with 2D Navier-Stokes equation 4Conclusion
In this paper, we introduced a novel method, that we named hybrid two-level MCMC approach, to compute the posterior mean of quantities of interest in Bayesian inverse problems. In this method, we take advantage of the fast evaluation of DL surrogates and of the high accuracy of numerical models. In particular, we have theoretically shown the potential to solve Bayesian inverse problems accurately up to an estimator error πͺ β ( β ) , by coupling one short MCMC chain generated by a high-fidelity numerical solver with mesh size β and another long MCMC chain generated with fast DL surrogates. We show the complete estimator error analysis and conclude that its theoretical bound is πͺ β ( 2 β πΏ ) with one long base MCMC chain of πͺ β ( 2 2 β πΏ ) number of DL surrogate samples and a short correction MCMC chain of πͺ ( ( 1 + 2 π ) 2 ) ) number of numerical samples, given the numerical forward model has an error rate of 2 β πΏ and DL-based forward surrogate has an error rate of 2 β πΏ + π . In addition, we show that with a surrogate speedup rate π of one forward solve, the overall speedup of our hybrid algorithm is πͺ β ( 2 2 β πΏ / max β‘ ( 1 π β 2 2 β πΏ , πΆ num πΆ DL β ( 1 + 2 π ) 2 ) ) . This implies that the overall speedup depends on the performance of the surrogate model, more speedup can be expected with high accurate surrogate models due to less numerical samples needed for error correction. To validate the theoretical findings, we performed numerical experiments on a Poisson equation, a nonlinear reaction-diffusion equation, a the Navier-Stokes equation. In all numerical experiments, we use both uniform priors and Gaussian priors. All results of our numerical experiments qualitatively validate our theoretical findings. However, we note that the theoretical result depends on the assumption of knowing the exact DL surrogate error, and the exact constants that appeared in the derivation of Theorem 2.2, which are rather challenging to obtain analytically. Due to this, we see in the numerical experiments that the actual optimal sample ratio from numerical sweeping test may deviate from the theoretical estimates of ( 1 + 2 π ) 2 / 2 2 β πΏ without taking into consideration of the constants πΆ num and πΆ DL . Yet, with a sweeping test, we show that for almost every experimental case, we can improve the accuracy of the MCMC estimator with an increasing number of numerical samples under our hybrid two-level MCMC method. We note that the theoretical framework proposed can also be used to understand the feasibility of using DL surrogates only (without hybridizing them with high-fidelity numerical solvers). More specifically, if the theoretical error estimates are already within the numerical error of a DL-only surrogate, then the use of the high-fidelity numerical solver will unlikely yield better accuracy. This theoretical result is particularly important, given the widespread use of DL surrogates in the field. As a final note, this paper focuses on a hybrid approach for the MCMC method to compute the posterior mean of quantities of interest in Bayesian inverse problems governed by PDEs; however, the same approach can in principle be applied to other Bayesian inverse problems not necessarily governed by PDEs, e.g. ODE governed Bayesian inverse problems, as well other methods such as filtering algorithms like the ensemble Kalman filter or sequential Monte Carlo methods.
References Appendix AHybrid two-level MCMC with Gaussian Prior
In Section 2.3, we introduced our new hybrid two-level MCMC for Bayesian inverse problems with uniform priors. However, in several instances, it may be convenient to work with Gaussian priors. For completeness, in this section we discuss the hybrid two-level MCMC method for Gaussian prior. Similar to the case of uniform priors, we consider a forward model that predicts the states π’ of a physical system given parameter π³ . In this case, the prior is Gaussian. Following the KL expansion (6) setup, we assume π := ( β π π β πΏ β β ( π· ) ) π
1 , 2 , β¦ , π β β 1 and π Β― := ( β π π β π 1 , β β ( π· ) ) π
1 , 2 , β¦ , π β β 1 . Then we define the measurable space ( π , Ξ π ) , with Ξ π := { π§ β β π , β π
1 π π π β | π§ π | < β } β β¬ β ( β π ) , and π is the parameter space. We denote the standard Gaussian measure in β by πΎ 1 . Hence, the prior can be defined as πΎ
β¨ π
1 π πΎ 1 on ( β π , β¬ β ( β π ) ) , and it completes the probability space ( π , Ξ , πΎ ) , noting that Ξ π has full Gaussian measure, i.e. πΎ β ( Ξ π )
1 .
Assumption A.1.
Let π’ be the solution of the the forward problem in equation (3). We assume that π’ β π , where π is a suitable vector spece, e.g., a Sobolev space. The FEM approximation gives
β π’ β ( π§ ) β π’ β β ( π§ ) β π β€ πΆ β exp β‘ ( π β β π
1 π π π β | π§ π | ) β ( 1 + β π
1 π π Β― π β | π§ π | ) β 2 β π .
(35)
The numerical error estimate in Assumption A.1 might not be true for all problems with Gaussian prior. It is problem-dependent. Yet, it is a typical error rate found in problems such as elliptic equations, diffusion problems, and parabolic problems with unknown coefficients, as investigated in [hoang2016convergence, MR4246090]. The right-hand side of Equation (35) differs from Assumption 2.2 made for a uniform prior, having an additional exponential term that depends on π³ .
This specific form of approximation error is of significant interest for our two-level hybrid MCMC approach, because the exponential term in the error will lead to divergence of the method (in contrast to the uniform prior introduced in Section 2.3). However, with Ferniqueβs theorem and an additional indication function to be presented below, we still can reach a similar posterior estimation error rate as the case in uniform prior. Using Assumption A.1, we can write the error between the DL-based surrogate and the numerical discretization as follows
β
π’
num
β
(
π§
)
β
π’
DL
β
(
π§
)
β
π
β€
πΆ
β
(
1
+
2
π
)
β
exp
β‘
(
π
β
β
π
1 π π π β | π§ π | ) β ( 1 + β π
1
π
π
Β―
π
β
|
π§
π
|
)
β
2
β
π
,
(36)
|
Ξ¦
num
β
(
π§
;
π¦
)
β
Ξ¦
DL
β
(
π§
;
π¦
)
|
β€
πΆ
β
(
1
+
2
π
)
β
exp
β‘
(
π
β
β
π
1 π π π β | π§ π | ) β ( 1 + β π
1 π π Β― π β | π§ π | ) β 2 β π .
(37)
Next we derive the two level MCMC approach for Gaussian prior. In order to avoid the unboundedness from the exponential term, we make use of the following switching function
π β ( π§ )
{ 1 , if β Ξ¦ num β ( π§ , π¦ ) β Ξ¦ DL β ( π§ , π¦ ) β€ 0 ,
0 , otherwise .
(38)
With the switching function (38), we can write the expected QoI as follows
( πΌ πΎ num β πΌ πΎ DL ) β [ π ]
1
π
num
β
β«
Ξ
π
exp
β‘
(
β
Ξ¦
num
)
β
π
β
π
β
(
π§
)
β
π
πΎ
β
1
π
DL
β
β«
Ξ
π
exp
β‘
(
β
Ξ¦
DL
)
β
π
β
π
β
(
π§
)
β
π
πΎ
+
1
π
num
β
β«
Ξ
π
exp
β‘
(
β
Ξ¦
num
)
β
π
β
(
1
β
π
β
(
π§
)
)
β
π
πΎ
β
1
π
DL
β
β«
Ξ
π
exp
β‘
(
β
Ξ¦
DL
)
β
π
β
(
1
β
π
β
(
π§
)
)
β
π
πΎ
1 π num β β« Ξ π exp β‘ ( β Ξ¦ num ) β ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π β π β ( π§ ) β π πΎ
( 1 π num β 1 π DL ) β β« Ξ π exp β‘ ( β Ξ¦ DL ) β π β π β ( π§ ) β π πΎ
1 π DL β β« Ξ π exp β‘ ( β Ξ¦ DL ) β ( exp β‘ ( Ξ¦ DL β Ξ¦ num ) β 1 ) β π β ( 1 β π β ( π§ ) ) β π πΎ
( 1 π num β 1 π DL ) β β« Ξ π exp β‘ ( β Ξ¦ num ) β π β ( 1 β π β ( π§ ) ) β π πΎ .
The constant ( 1 / π num β 1 / π DL ) , can be estimated via
( 1 π num β 1 π DL )
= 1 π num β π DL β β« Ξ π ( exp β‘ ( β Ξ¦ DL β ( π§ , π¦ ) ) β exp β‘ ( β Ξ¦ num β ( π§ , π¦ ) ) ) β ( π β ( π§ ) + 1 β π β ( π§ ) ) β π πΎ β ( π§ )
1
π
num
β
π
DL
β
β«
Ξ
π
exp
β‘
(
β
Ξ¦
num
β
(
π§
,
π¦
)
)
β
(
exp
β‘
(
Ξ¦
num
β
(
π§
,
π¦
)
β
Ξ¦
DL
β
(
π§
,
π¦
)
)
β
1
)
β
π
β
(
π§
)
β
π
πΎ
β
(
π§
)
+
1
π
num
β
π
DL
β
β«
Ξ
π
exp
β‘
(
β
Ξ¦
DL
β
(
π§
,
π¦
)
)
β
(
1
β
exp
β‘
(
Ξ¦
DL
β
(
π§
,
π¦
)
β
Ξ¦
num
β
(
π§
,
π¦
)
)
)
β
(
1
β
π
β
(
π§
)
)
β
π
πΎ
β
(
π§
)
1 π DL β πΌ πΎ num β [ ( exp β‘ ( Ξ¦ num β ( π§ , π¦ ) β Ξ¦ DL β ( π§ , π¦ ) ) β 1 ) β π β ( π§ ) ]
- 1 π num β πΌ πΎ DL β [ ( 1 β exp β‘ ( Ξ¦ DL β ( π§ , π¦ ) β Ξ¦ num β ( π§ , π¦ ) ) ) β ( 1 β π β ( π§ ) ) ] .
Combining equations (2.3), (A) and (A), we can derive the overall estimator of our hybrid two-level MCMC approach
π β β π¦ β π β π β π β π β ( π )
π πΎ num β [ π΄ 1 ] + π πΎ num β [ π΄ 3 ] β π πΎ DL β [ π΄ 4 + π΄ 8 ]
- π πΎ DL β [ π΄ 2 ]
- π πΎ DL β [ π΄ 5 ] β π πΎ num β [ π΄ 6
- π΄ 7 ]
- π πΎ DL β [ π ] ,
where the terms π΄ 1 , π΄ 2 , π΄ 3 , π΄ 4 , π΄ 5 , π΄ 6 , π΄ 7 and π΄ 8 are defined as follows
π΄ 1
(
1
β
exp
β‘
(
Ξ¦
num
β
(
π§
)
β
Ξ¦
DL
β
(
π§
)
)
)
β
π
β
(
π§
)
β
π
β
(
π§
)
,
π΄
2
(
exp
β‘
(
Ξ¦
DL
β
(
π§
)
β
Ξ¦
num
β
(
π§
)
)
β
1
)
β
π
β
(
π§
)
β
(
1
β
π
β
(
π§
)
)
,
π΄
3
(
exp
β‘
(
Ξ¦
num
β
(
π§
)
β
Ξ¦
DL
β
(
π§
)
)
β
1
)
β
π
β
(
π§
)
,
π΄
4
π
β
(
π§
)
β
π
β
(
π§
)
,
π΄
5
(
1
β
exp
β‘
(
Ξ¦
DL
β
(
π§
)
β
Ξ¦
num
β
(
π§
)
)
)
β
(
1
β
π
β
(
π§
)
)
π΄
6
exp
β‘
(
Ξ¦
num
β
Ξ¦
DL
)
β
π
β
(
π§
)
β
π
β
(
π§
)
,
π΄
7
π
β
(
π§
)
β
(
1
β
π
β
(
π§
)
)
,
π΄
8
exp β‘ ( Ξ¦ DL β ( π§ ) β Ξ¦ num β ( π§ ) ) β π β ( π§ ) β ( 1 β π β ( π§ ) ) .
We now perform the error analysis of our method, under the assumption of Gaussian priors. In analogy with what we have done for uniform priors in Section 2.3, we decompose the error in three terms:
πΌ πΎ π¦ β [ π ] β π hybrid β [ π ]
I + II + III ,
(41a)
I := πΌ πΎ π¦ β [ π ] β πΌ πΎ num β [ π ] ,
(41b)
II := πΌ πΎ π· β πΏ β [ π ] β π π num πΎ π· β πΏ β [ π ] ,
(41c)
III := πΌ πΎ num β [ π΄ 1 ] β π π num πΎ num β [ π΄ 1 ] + πΌ πΎ num β [ π΄ 3 ] β πΌ πΎ DL β [ π΄ 4 + π΄ 8 ]
β π π num πΎ num β [ π΄ 3 ] β π π num πΎ DL β [ π΄ 4 + π΄ 8 ] + πΌ πΎ DL β [ π΄ 2 ] β π π num πΎ DL β [ π΄ 2 ]
- πΌ πΎ DL β [ π΄ 5 ] β πΌ πΎ num β [ π΄ 6
- π΄ 7 ] β π π num πΎ DL β [ π΄ 5 ] β π π num πΎ num β [ π΄ 6
- π΄ 7 ] .
(41d)
Similarly to Section 2.3, for each error term, we have the following error bounds
| I | < πΆ β 2 β πΏ ,
(42a)
| II | < π DL β 1 / 2 ,
(42b)
β° β [ | III | 2 ] < πΆ β ( 1 + 2 π ) 2 β π num β 1 β 2 β 2 β πΏ .
(42c)
Therefore, choosing π DL
πΆ DL β 2 2 β πΏ and π num
πΆ num β ( 1 + 2 π ) 2 , allows us to obtain a theorem for the overall error estimate.
Theorem A.3.
With π DL
πΆ DL β 2 2 β πΏ and π num
πΆ num β ( 1 + 2 π ) 2 , we have the following theoretical error estimate of our hybrid two-level MCMC approach under Gaussian priors
β° hybrid β [ | πΌ πΎ π¦ β [ π ] β π hybrid β [ π ] | ] β€ πΆ hybrid β 2 β πΏ .
(43)
From Theorem A.3, we see the same results as the one from Theorem 2.2 derived from uniform prior setup. The theorem and the preceding assumptions are typically valid for log-normal priors with elliptic, diffusion, and parabolic equations. The proof for the two-dimensional Navier-Stokes equation is not available to the best of our knowledge. However, some experimental results also show the theorem for multilevel MCMC with Gaussian prior works for the two-dimensional Navier-Stokes equation [MR4523340].
Appendix BForward numerical solvers and deep learning surrogate models
We present the details of the forward numerical solvers used for solving each of the problems presented in Section 3, along with the corresponding DL surrogate.
B.1Poisson equation
The Poisson equation 30 in Section 3.1 is solved with the finite element package Fenicsx [barrata2023dolfinx], whereby first-order Lagrange finite elements are used to discretize the equation. The mesh adopted is constituted of 32 Γ 32 elements as shown in Figure 2a. A direct LU solver from MUMPS backend is used to solve the assembled linear system on CPU.
For the simple uniform prior setup in Section 3.1.1, we choose a fully connected ReLU neural network to learn the forward mapping from the 4000 generated training samples. The input field of 33 Γ 33 is flattened and fed in as input data. Two hidden layers are included, each with 512 nodes. The neural network is implemented with PyTorch. The trained neural network is used as the DL-based surrogate model in the experiment.
For the Gaussian prior setup in Section 3.1.2, we choose a convolutional neural network (CNN) to be our DL-based surrogate model. The CNN model consists of 3 encoding layers, 1 fully connected layer, and 3 decoding layers. There are 8 kernels in each convolutional layer, and the kernel size is ( 3 , 3 ) with stride size ( 2 , 2 ) .
B.2Nonlinear reaction-diffusion equation
To solve the reaction diffusion equation (32) in Section 3.2, we also use the finite element package Fenicsx [barrata2023dolfinx], whereby first-order Lagrange finite elements are used to discretize the equation. The resulting nonlinear system is solved with the Newton solver from the PETSC backend. Each linearized Newton iteration step is solved with LU direct solver with MUMPS backend.
For the DL-based surrogate model, we choose message passing graph neural network (MPGNN) for the uniform prior case. We refer to the Graph-PDE architecture [iakovlevlearning] as our reference. Instead of training the MPGNN for a time dependent problem, here we train the neural network for a time independent problem. Two fully connected multilayer neural networks are used for the message passing and state update. There are two hidden layers each with 64 nodes in both the message passing and state update neural networks. The hidden state vector output from the message passing neural network is of size 64. A Dirichlet boundary condition is also imposed on the MPGNN. Five layers of MPGNN are stacked in the model used in the experiments. The MPGNN is trained with 2000 training samples with Adam optimizer and trained for 10000 epoches.
We then choose U-net proposed in [10.1007/978-3-319-24574-4_28] for the Gaussian prior experiment. It is well-known for its outstanding performance in multi-scale physical problems. The U-net consists of 3 layers, each with dimensions of 32 Γ 32 , 16 Γ 16 , and 8 Γ 8 . For each convolutional layer we have 8 kernel with size ( 3 , 3 ) and stride size ( 1 , 1 ) .
B.3Navier-Stokes equations in the vorticity form
To solve the Navier-Stokes equations 33 in Section 3.3, we coded a simple numerical pseudo-spectral solver with PyTorch, which is accelerated on GPU. A total of 64 Γ 64 collocation points are used for the experiments. The interested reader can refer to [peyret2002spectral] for details of the spectral method implemented.
For the DL-based surrogate model, we first choose the DeepONet [lu2021learning] as the deep learning model for the uniform prior experiment. Two fully connected multilayer neural networks are used as the branch net and trunk net. Both the branch net and trunk net have two hidden layers with 64 nodes in the neural network. The DeepONet model is trained with 4000 samples of numerical data for 10000 epoches.
Then we choose the Fourier Neural Operator for the Gaussian prior experiment, as it demonstrated its ability to learn the dynamics of the Navier-Stokes equation in [DBLP:conf/iclr/LiKALBSA21] and has also been shown to be used in a Bayesian inversion problem setup with MCMC. Specifically, the two-dimensional Fourier neural operator with tensor layers is used. In this experiment we used 12 modes for height and width, 8 hidden channels, and 4 layers in the FNO. The neural network is trained with 2000 training data for 10000 epochs.
Appendix CDiagnostics details of numerical experiment 3.1.1 Figure 13:Left) Trace plots of MCMC chains with DL-based surrogate models from eight experiments with different initial sample showing good mixing of samples. Right) Autocorrelation Function(ACF) plot from the same eight MCMC chains showing rapid decay, which indicates good mixing and a high number of Effective Sample Size (ESS). Figure 14:Histogram of MCMC chains with DL-based surrogate models from eight experiments with different initial sample showing similar posterior distribution of the samples with good mixing (a)Sample mean of eight MCMC chains (b)Potential Scale Reduction Factor Figure 15:Left) Cumulative average of samples of eight independent DL-based MCMC chain showing good inter-chain convergence. Right) PSRF of eight independent DL-based surrogate MCMC chains showing rapid decay and stable small PSRF values indicating good inter-chain convergence Figure 16:Left) Trace plot of QoI ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π§ from 1000 numerical samples (1% of total DL-based surrogate samples) from eight independent experiments with different initial sample showing good mixing of samples. Right) Autocorrelation Function(ACF) plot from the same eight MCMC chains showing rapid decay indicating good mixing and high number of Effective Sample Size(ESS) Figure 17:Histogram of of QoI ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π§ from 1000 numerical samples (1% of total DL-based surrogate samples) from eight independent MCMC chain with different initial sample showing similar posterior distribution of QoI with good mixing Figure 18:Left) Trace plot of QoI exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 from 1000 numerical samples (1% of total DL-based surrogate samples) from eight independent MCMC chain with different initial sample showing good mixing of samples. Right) Autocorrelation Function (ACF) plot from the same eight MCMC chains showing rapid decay indicating good mixing and high number of Effective Sample Size(ESS) Figure 19:Histogram of QoI ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π§ from 1000 numerical samples (1% of total DL-based surrogate samples) from eight independent experiments with different initial sample showing similar posterior distribution of the QoI with good mixing. (a)Sample mean of eight MCMC chains (b)Potential Scale Reduction Factor Figure 20:Left) Cumulative average of QoI ( 1 β exp β‘ ( Ξ¦ num β Ξ¦ DL ) ) β π§ from eight independent numerical MCMC chains showing good inter-chain convergence. Right) PSRF of the same eight MCMC chains showing rapid decay and stable small PSRF values indicating a good inter-chain convergence. (a)Sample mean of eight MCMC chains (b)Potential Scale Reduction Factor Figure 21:Left) Cumulative average of QoI exp β‘ ( Ξ¦ num β Ξ¦ DL ) β 1 from eight independent numerical MCMC chains with different initial sample showing good inter-chain convergence. Right) PSRF of the same eight MCMC chains showing rapid decay and stable small PSRF values indicating a good inter-chain convergence Generated on Sun Oct 19 08:48:54 2025 by LaTeXML Report Issue Report Issue for Selection
Xet Storage Details
- Size:
- 105 kB
- Xet hash:
- 2cf1538c9375908c06f68050c6c9eeed7e74e6ee5c313a34bd6a58c6c86418c1
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.