Elsevier

Advances in Water Resources

Volume 102, April 2017, Pages 161-177
Advances in Water Resources

Efficient geostatistical inversion of transient groundwater flow using preconditioned nonlinear conjugate gradients

https://doi.org/10.1016/j.advwatres.2016.12.006Get rights and content

Highlights

  • We present a preconditioned conjugate gradient method for geostatistical inversion.

  • The prior covariance matrix is used as preconditioner at negative computational cost.

  • The approach incorporates linearized uncertainty quantification.

  • The method is particularly efficient for the inversion of transient data.

  • Transient inversion may speed up experiments and improve quality of inversion results.

Abstract

In the geostatistical inverse problem of subsurface hydrology, continuous hydraulic parameter fields, in most cases hydraulic conductivity, are estimated from measurements of dependent variables, such as hydraulic heads, under the assumption that the parameter fields are autocorrelated random space functions. Upon discretization, the continuous fields become large parameter vectors with O(104107) elements. While cokriging-like inversion methods have been shown to be efficient for highly resolved parameter fields when the number of measurements is small, they require the calculation of the sensitivity of each measurement with respect to all parameters, which may become prohibitive with large sets of measured data such as those arising from transient groundwater flow.

We present a Preconditioned Conjugate Gradient method for the geostatistical inverse problem, in which a single adjoint equation needs to be solved to obtain the gradient of the objective function. Using the autocovariance matrix of the parameters as preconditioning matrix, expensive multiplications with its inverse can be avoided, and the number of iterations is significantly reduced. We use a randomized spectral decomposition of the posterior covariance matrix of the parameters to perform a linearized uncertainty quantification of the parameter estimate.

The feasibility of the method is tested by virtual examples of head observations in steady-state and transient groundwater flow. These synthetic tests demonstrate that transient data can reduce both parameter uncertainty and time spent conducting experiments, while the presented methods are able to handle the resulting large number of measurements.

Introduction

The spatially distributed hydraulic conductivity K(x) [LT1] is widely considered as the most important parameter field for groundwater flow and, indirectly, solute transport. Estimating K(x) is a prerequisite for groundwater management, assessment of solute transport in aquifers, and remediation of contaminated sites. Because direct measurement of hydraulic conductivity can only be done in the laboratory using samples of aquifer material, the estimation of K(x) in the field mostly relies on hydraulic-head measurements, either under conditions of ambient groundwater flow, or during hydraulic tests (Liu, Yeh, Gardiner, 2002, Liu, Illman, Craig, Zhu, Yeh, 2007, Straface, Yeh, Zhu, Troisi, Lee, 2007, Zhu, Yeh, 2005). Additional information about hydraulic conductivity can be obtained from concentration data, e.g., from tracer tests, due to the dependence of the velocity on K(x) (Datta-Gupta, Yoon, Vasco, Pope, 2002, Vasco, Datta-Gupta, 1999, Yeh, Zhu, 2007). Traditional methods of estimating K(x) rely on analytical solutions of the groundwater-flow and potentially solute-transport equations, typically assuming uniform parameter values (e.g., Kruseman, Ridder, et al., 1990, Rumynin, 2011). In hydrogeological and engineering practice, numerical models are commonly calibrated to fit measured hydraulic heads and other observations, conceptualizing the subsurface as a set of space-filling, simple-shaped geometric bodies, such as layers and blocks, with uniform hydraulic properties (e.g., Hill and Tiedemann, 2007). Identification of the hydraulically relevant geobodies, however, remains a challenge, and the hydraulic variability within the geological units may be larger than between them.

In the last three decades, geostatistical methods have become widely accepted in order to characterize the spatial variability of hydraulic subsurface properties across many scales, notwithstanding the identification of dominant geological features (e.g., Rubin, 2003; Kitanidis, 1997; Zhang, 2002). The key assumption is that the hydraulic-conductivity field is the outcome of a spatially autocorrelated random process that is considered stationary unless data strongly supports spatial variation of statistical properties. The geostatistical premise can also be used as prior information in the estimation of hydraulic conductivity as a continuous spatial field from direct data, i.e., conductivity values themselves, and indirect data such as hydraulic-head measurements. Upon discretization, one hydraulic-conductivity value is assigned to each node or element of a numerical grid, leading to O(104) parameter values in 2-D, and O(106107) in 3-D applications. In this context, the geostatistical prior information acts as a regularization term, because the number of measurements typically is by orders of magnitude smaller than the number of grid cells, resulting in an ill-posed inverse problem.

Most geostatistical inverse methods assume that the logarithm of hydraulic conductivity Y(x) is an – at least blockwise – second-order stationary multi-Gaussian field. Some approaches replace the measurements of indirect data, such as hydraulic heads, by virtual measurements of log-hydraulic conductivity itself at so-called pilot points (Doherty, 2003, RamaRao, LaVenue, De Marsily, Marietta, 1995), master points (Gómez-Hernánez et al., 1997), or anchors (Rubin et al., 2010). Then the log-hydraulic conductivity field between these points is estimated by interpolation or conditional simulation, and the inversion becomes estimating the virtual log-conductivity measurements. This approach effectively reduces the number of parameters to be estimated, but can lead to bias in the conditional uncertainty analysis, because the spatial patterns of the conditional Y(x)-field and its uncertainty differ whether hydraulic-head or log-conductivity measurements are used for conditioning Kitanidis (1998).

As alternatives to reduce the dimension of the parameter space, the eigen-decomposition-based Karhunen–Loève expansion (e.g., Li and Cirpka, 2006; Li and Zhang, 2007) and truncated singular-value decomposition methods (Kitanidis, Lee, 2014, Lee, Kitanidis, 2014) have been used in geostatistical inversion. These techniques act as low-pass filters. Deciding upon the right number of terms requires balancing the computational effort and the accuracy of the inversion.

An extended group of mathematically similar geostatistical inverse methods rely on variants of the cokriging equations, which can be derived from the Gauss–Newton method applied to the geostatistical inverse problem (Hoeksema, Kitanidis, 1984, Kitanidis, 1995, Rubin, Dagan, 1987, Valstar, McLaughlin, Te Stroet, van Geer, 2004, Yeh, Gutjahr, Jin, 1995, Yeh, Jin, Hanna, 1996). While in the original Gauss–Newton method, the order of the system of equations to be solved is the number of parameters (McLaughlin and Townley, 1996), it is reduced to the number of measurements plus the number of deterministic drift coefficients in the cokriging equations. A benefit of the approach is that it directly yields a linearized uncertainty estimate. A disadvantage is that it requires the calculation of the full sensitivity matrix, that is the sensitivity of each measurement with respect to each parameter. This can be achieved by analytical linearization about the prior mean (Hoeksema, Kitanidis, 1984, Rubin, Dagan, 1987), thus neglecting the non-linear relationship between hydraulic heads and log-conductivity in the inversion, direct numerical differentiation, or application of adjoint-state methods (Sun and Yeh, 1990). The latter two approaches require solving equations of the same complexity as the forward problem as often as either the number of parameters or the number of measurements. That is, the non-linear cokriging-type inversions become computational prohibitive when the log-conductivity field is highly resolved and many measurements are to be inverted.

A large number of measurements is typically achieved in monitoring of transient state variables, such as hydraulic heads during transient pumping tests or concentrations during tracer tests, which may pose a problem to cokriging-like geostatistcal inversion as discussed above. Some authors recommend picking a selected small number of observations from the continuous data streams (Cardiff, Barrash, 2011, Cardiff, Barrash, Kitanidis, 2012, Liu, Illman, Craig, Zhu, Yeh, 2007), whereas others condensate the information to a few temporal moments (Cirpka, Kitanidis, 2000, Leube, Nowak, Schneider, 2012, Li, Nowak, Cirpka, 2005, Pollock, Cirpka, 2010, Zhu, Yeh, 2006). Both approaches are associated with loosing information.

Recently, Ensemble-Kalman methods have become popular in geostatistical inversion (Chen, Zhang, 2006, Franssen, Kinzelbach, 2008, Franssen, Kinzelbach, 2009, Nowak, 2009, Schöniger, Nowak, Hendricks Franssen, 2012). In these methods, an ensemble of realizations is used to compute the covariance matrices involving the dependent variables. By this, the computation of the sensitivity matrix is circumvented. However, to avoid bias in the uncertainty estimates, denoted filter inbreeding, either a large number of realizations are needed or techniques to inflate the conditional covariance have to be applied after each conditioning step. As an alternative to performing Monte–Carlo simulations, second-order moment-generating equations have been suggested to compute the underlying covariance matrices, but here the number of partial differential equations to be solved equals the number of grid elements (Panzeri, Riva, Guadagnini, Neuman, 2013, Panzeri, Riva, Guadagnini, Neuman, 2014).

In summary, the existing geostatistical inverse methods either rely on parameterizing the continuous fields with a few terms, require the evaluation of as many adjoint problems as the number of measurements, condensate the data, or demand the evaluation of large ensembles of parameter fields. Thus, despite progress made in inversion methods in the last three decades, a choice between accuracy and computational efficiency has to be made.

The geostatistical inverse problem for transient data sets is conceptually similar to large-scale data-assimilation problems, such as those occurring in weather forecast (Caya, Sun, Snyder, 2005, Marécal, Mahfouf, 2002, Županski, Mesinger, 1995) and oceanography (Weaver et al., 2003). Here, very large vectors of state variables (with up to O(108) elements) are updated using smaller, but still large vectors of observations (with up to O(106) elements) that depend on the state variables in nonlinear ways. In the variational approaches toward data assimilation, an objective function is minimized that resembles that of the geostatistical inverse problem discussed here. The sheer size of the state and observation vectors prohibits methods involving the full sensitivity matrix. Conjugate Gradient methods, by contrast, require only calculating the sensitivity of the single-valued objective function with respect to the vector of values to be estimated. The latter task can efficiently be done by adjoint-state methods. The method can be expedited by choosing an appropriate preconditioning matrix.

Conjugate Gradient methods have already been used in the inversion of groundwater data, among others, in the pilot-point/master-point approaches (Gómez-Hernánez, Sahuquillo, Capilla, 1997, RamaRao, LaVenue, De Marsily, Marietta, 1995); but so far they have not been applied to the full geostatistical problem without additional parameterization. A potential reason for that could be that evaluating the gradient of the prior term requires solving a large system of equations involving the dense prior covariance matrix on the left-hand side. In the present paper, we will show that the latter can be avoided by using the prior covariance matrix for preconditioning of the Conjugate Gradient method.

The remaining article is structured as follows: Section 2 describes the inverse problem that is to be solved, stating the forward model, its governing equations and the derivation of an objective function for the inverse problem through Bayesian statistics. Section 3 discusses the solution of the arising optimization problem, describes the Gauss–Newton algorithm as a reference scheme that is often applied and introduces the Conjugate Gradient method preconditioned with the prior covariance matrix as an alternative. Section 4 lists the software packages used for the numerical implementation and specifies implementation details. Section 5 shows the application of the discussed methods in a synthetic test case, comparing the schemes regarding their computational cost and accuracy. Finally, Section 6 introduces an eigen-decomposition-based algorithm for linearized uncertainty quantification and applies it to the aforementioned synthetic test case.

Section snippets

Governing equations

For simplicity, we restrict the following analysis to the inference of the log-hydraulic conductivity field Y(x) from the measurements of hydraulic heads ϕ(tm, xm) [L] at given measurement locations xm [L] and times tm [T]. In 3-D domains, the hydraulic-head field meets the following partial differential equation (PDE), known as groundwater-flow equation: Ssϕt·(Kϕ)=qϕ,in which Ss=θϱϱϕ+θϕ[L1] denotes the specific storativity, where θ [-] is the porosity and ϱ [ML3] is the mass density

Reference: Gauss–Newton method

Minimization of L can be achieved using a variety of methods, e.g., the Gauss–Newton method or its stabilized version, the Levenberg–Marquardt method. In addition to the gradient of L, these methods require information about the Hessian of L and therefore typically assemble the whole sensitivity matrix HYz.

The standard method to compute HYz is via adjoint-state theory, both due to its efficiency and its accuracy Michalak and Kitanidis (2004). Assembly of the whole matrix needs the solution of mz

Numerical implementation

The method described above is realized in the framework of the open-source software Distributed and Unified Numerics Environment (DUNE) and its module DUNE-PDELab. This software package provides interfaces to a large variety of computational grids (Bastian, Blatt, Dedner, Engwer, Klöfkorn, Kornhuber, Ohlberger, Sander, 2008a, Bastian, Blatt, Dedner, Engwer, Klöfkorn, Ohlberger, Sander, 2008b), finite-element basis functions and efficient solvers for PDEs (Blatt, Bastian, 2007, Blatt, Bastian,

Numerical examples

We compare numerical implementations of the three methods described above, namely: (a) the original CG scheme, Eq. (17), (b) the preconditioned CG scheme, Eq. (20), and as reference (c) the Gauss–Newton scheme, Eqs. (15), using the code of Schwede et al. (2012), which was implemented in the same framework.

In the applications, we consider a domain of size 100 m ×  100 m with an injection well in the upper part and an extraction well in the lower part of the domain. As illustrated in Fig. 1, an

Uncertainty quantification

The PCG scheme presented above does not yield information about the uncertainty of the estimate, in contrast to methods that assemble the full sensitivity matrix HYz, like the Gauss–Newton method, which automatically provides linearized uncertainty quantification. This makes a postprocessing step for uncertainty quantification necessary, since the uncertainty of the parameters has to be taken into account when they are used for predictions. This section closely follows an approach detailed in

Conclusions

We have presented a geostatistical inversion method that can deal with large data sets, as generated by time-dependent groundwater flow. The inversion kernel is a Preconditioned Conjugate Gradient method using the prior autocovariance matrix of the parameters as preconditioner. The gradient of the objective function is computed by an adjoint formulation, and the computational effort for this task does not depend on the number of measurements to be inverted.

We have demonstrated the feasibility

Acknowledgments

This study has been funded by the Baden-Württemberg Stiftung in its high-performance computing program, contract HPC-8, and by the German Federal Ministry for Education and Research (BMBF) within the geotechnologies program under the grant 03G0742A.

References (70)

  • E. Allen et al.

    Numerical approximation of the product of the square root of a matrix with a vector

    Linear Algebra Appl.

    (2000)
  • P. Bastian et al.

    A generic grid interface for parallel and adaptive scientific computing. Part II: implementation and tests in dune

    Computing

    (2008)
  • P. Bastian et al.

    A generic grid interface for parallel and adaptive scientific computing. Part I: abstract framework

    Computing

    (2008)
  • P. Bastian et al.

    Generic implementation of finite element methods in the distributed and unified numerics environment (dune)

    Kybernetika

    (2010)
  • M. Blatt et al.

    The iterative solver template library

  • M. Blatt et al.

    On the generic parallelisation of iterative solvers for the finite element method

    Int. J. Comput. Sci. Eng.

    (2008)
  • T. Bui-Thanh et al.

    Extreme-scale UQ for Bayesian inverse problems governed by PDEs

    Proceedings of the International Conference on High Performance Computing, Networking, Storage and Analysis

    (2012)
  • M. Cardiff et al.

    3-D transient hydraulic tomography in unconfined aquifers with fast drainage response

    Water Resour. Res.

    (2011)
  • M. Cardiff et al.

    A field proof-of-concept of aquifer imaging using 3-d transient hydraulic tomography with modular, temporarily-emplaced equipment

    Water Resour. Res.

    (2012)
  • J. Carrera et al.

    Estimation of aquifer parameters under transient and steady state conditions: 3. application to synthetic and field data

    Water Resour. Res.

    (1986)
  • A. Caya et al.

    A comparison between the 4DVAR and the ensemble Kalman filter techniques for radar data assimilation

    Mon. Weather Rev.

    (2005)
  • Y. Chen et al.

    Data assimilation for transient flow in geologic formations via ensemble Kalman filter

    Adv. Water Resour.

    (2006)
  • O.A. Cirpka et al.

    Sensitivity of temporal moments calculated by the adjoint-state method and joint inversing of head and tracer data

    Adv. Water Resour.

    (2000)
  • A. Datta-Gupta et al.

    Inverse modeling of partitioning interwell tracer tests: a streamline approach

    Water Resour. Res.

    (2002)
  • C.R. Dietrich et al.

    Fast and exact simulation of stationary gaussian processes through circulant embedding of the covariance matrix

    SIAM J. Sci. Comput.

    (1997)
  • J. Doherty

    Ground water model calibration using pilot points and regularization

    Groundwater

    (2003)
  • H.H. Franssen et al.

    Real-time groundwater flow modeling with the ensemble Kalman filter: joint estimation of states and parameters and the filter inbreeding problem

    Water Resour. Res.

    (2008)
  • H.H. Franssen et al.

    Ensemble Kalman filtering versus sequential self-calibration for inverse modelling of dynamic groundwater flow systems

    J. Hydrol.

    (2009)
  • M. Frigo et al.

    The design and implementation of fftw3

    Proc. IEEE

    (2005)
  • J.J. Gómez-Hernánez et al.

    Stochastic simulation of transmissivity fields conditional to both transmissivity and piezometric data – I. theory

    J. Hydrol.

    (1997)
  • G. Guennebaud et al.

    Eigen v3

    (2010)
  • W.W. Hager

    Updating the inverse of a matrix

    SIAM Rev.

    (1989)
  • N. Halko et al.

    Finding structure with randomness: probabilistic algorithms for constructing approximate matrix decompositions

    SIAM Rev.

    (2011)
  • M. Hill et al.

    Effective Groundwater Model Calibration

    (2007)
  • R.J. Hoeksema et al.

    An application of the geostatistical approach to the inverse problem in two-dimensional groundwater modeling

    Water Resour. Res.

    (1984)
  • D. Kalman

    A singularly valuable decomposition: the SVD of a matrix

    Coll. Math. J.

    (1996)
  • P.K. Kitanidis

    Quasi-linear geostatistical theory for inversing

    Water Resour. Res.

    (1995)
  • P.K. Kitanidis

    Introduction to Geostatistics: Applications in Hydrogeology

    (1997)
  • P.K. Kitanidis

    How observations and structure affect the geostatistical solution to the steady-state inverse problem

    Groundwater

    (1998)
  • P.K. Kitanidis et al.

    Principal component geostatistical approach for large-dimensional inverse problems

    Water Resour. Res.

    (2014)
  • O. Klein

    Preconditioned and Randomized Methods for Efficient Bayesian Inversion of Large Data Sets and their Application to Flow and Transport in Porous Media

    (2016)
  • G.P. Kruseman et al.

    Analysis and Evaluation of Pumping Test Data

    (1990)
  • LeeJ. et al.

    Large-scale hydraulic tomography and joint inversion of head and tracer data using the principal component geostatistical approach (PCGA)

    Water Resour. Res.

    (2014)
  • P. Leube et al.

    Temporal moments revisited: why there is no better way for physically based model reduction in time

    Water Resour. Res.

    (2012)
  • H. Li et al.

    Probabilistic collocation method for flow in porous media: comparisons with other stochastic methods

    Water Resour. Res.

    (2007)
  • Cited by (14)

    • Mapping conduits in two-dimensional heterogeneous karst aquifers using hydraulic tomography

      2023, Journal of Hydrology
      Citation Excerpt :

      By interpreting the head responses and assessing the associated uncertainties, different kinds of HT inversion schemes have been reported in literature to determine the spatial distribution of hydraulic parameters. These inverse methods include general geostatistical inverse methods (Yeh and Liu, 2000; Zhu and Yeh, 2005; Li and Cirpka, 2006; Klein et al., 2017; Schwede et al., 2014; Cardiff et al., 2009, 2013; Mao et al., 2013a, 2013b; Zhao et al., 2015; Zhao and Illman, 2018; Li et al., 2021), travel time-based inverse methods (Brauchler et al., 2003, 2010, 2013; Vasco et al., 2019), pilot-point-based approaches (Lavenue and De Marsily, 2001; Castagna et al. 2011), sparse nonlinear optimizer and stochastic Newton approach (Wang et al., 2016, 2017), cellular automata-based approaches (Fischer et al., 2017a,b), and discrete network deterministic inversion method (Fischer et al., 2018, 2020). In this study, we focus on the geostatistics-based and travel time-based inversion methods of HT.

    • High-dimensional inverse modeling of hydraulic tomography by physics informed neural network (HT-PINN)

      2023, Journal of Hydrology
      Citation Excerpt :

      The major challenge faced by GA is its applicability when the dimension of the target variable is large for estimating a high-resolution parameter field. The cost of computing and storing the associated full-rank covariance matrix is very high, especially when using gradient-based methods such as Newton’s method to evaluate the Jacobian matrix at each iteration (Ambikasaran et al., 2013; Klein et al., 2017; Liu and Kitanidis, 2011; Liu et al., 2013; Obiefuna and Eslamian, 2019). Many efforts have been made to optimize GA for large-scale inverse problems from two aspects: dimensionality reduction and efficient numerical methods.

    • Bayesian inverse modeling of large-scale spatial fields on iteratively corrected principal components

      2021, Advances in Water Resources
      Citation Excerpt :

      For high-dimensional inverse problems with millions of unknowns, major computational costs include storage and computations of covariance matrices, and iterative implementation of forward models to determine the Jacobian matrix for nonlinear problems. In the past decades, many computational advances have been proposed to facilitate the storage and computation of covariance matrices, and to reduce the number of forward model computations within the classic Bayesian geostatistical framework for solving large-dimensional inverse problems (Nowak et al., 2003; Nowak and Cirpka, 2004; Liu and Kitanidis, 2011; Saibaba et al., 2012; Ambikasaran et al., 2013; Liu et al., 2013; Kitanidis and Lee, 2014; Lee and Kitanidis, 2014; Lee et al., 2016; Lin et al., 2017; Klein et al., 2017). Recently, we developed a reformulated framework of Bayesian inverse modeling for large-scale spatial fields on reduced dimensions comprised by principal components (Zhao and Luo, 2020).

    • Efficient 3D inversions using the Richards equation

      2018, Computers and Geosciences
      Citation Excerpt :

      Alternatively, if prior information can be formulated using a statistical framework, we can use Bayesian techniques to obtain an estimator through the Maximum A Posteriori model (MAP) (Kaipio and Somersalo, 2004). In the context of Bayesian estimation, gradient based methods are also important, as they can be used to efficiently sample the posterior (Bui-Thanh and Ghattas, 2015; Liu et al., 2017; Klein et al., 2017). A number of authors have sought solutions for the inverse problem, where the forward problem is the Richards equation (cf. Bitterlich and Knabner (2002); Iden and Durner (2007); Šimunek and Senja (2012) and references within).

    View all citing articles on Scopus
    View full text