Sparse polynomial chaos expansion based on Bregman-iterative greedy coordinate descent for global sensitivity analysis

https://doi.org/10.1016/j.ymssp.2021.107727Get rights and content

Highlights

  • A novel algorithm is proposed to build sparse PCE for global sensitivity analysis.

  • The accuracy and efficiency are assessed on benchmarks and engineering examples.

  • A detailed comparison is made with OMP, LAR and two adaptive algorithms.

  • The algorithm provides a better tradeoff among accuracy, complexity and efficiency.

Abstract

Polynomial chaos expansion (PCE) is widely used in a variety of engineering fields for uncertainty and sensitivity analyses. The computational cost of full PCE is unaffordable due to the ‘curse of dimensionality’ of the expansion coefficients. In this paper, a novel methodology for developing sparse PCE is proposed by making use of the efficiency of greedy coordinate descent (GCD) in sparsity exploitation and the capability of Bregman iteration in accuracy enhancement. By minimizing an objective function composed of the l1 norm (sparsity) of the polynomial chaos (PC) coefficients and regularized l2 norm of the approximation fitness, the proposed algorithm screens the significant basis polynomials and builds an optimal sparse PCE with model evaluations much fewer than unknown coefficients. To validate the effectiveness of the developed algorithm, several benchmark examples are investigated for global sensitivity analysis (GSA). A detailed comparison is made with the well-established orthogonal matching pursuit (OMP), least angle regression (LAR) and two adaptive algorithms. Results show that the proposed method is superior to the benchmark methods in terms of accuracy while maintaining a better balance among accuracy, complexity and computational efficiency.

Introduction

Along with the ever-increasing complexity of computer models for engineering simulations, the inherent uncertainties of input data and model parameters are evolving rapidly. In this context, characterizing the uncertainties within a computer model is of great importance, and this has motivated the development of a variety of numerical techniques for the emerging field of uncertainty quantification. To efficiently quantify the effect of variation in each input parameter on model outputs, one popular technique is to substitute a computationally expensive model with a surrogate model that possesses similar quantities of interest such as statistical moments and the distribution of model outputs.

Surrogate model (also known as metamodel) is a mathematical or numerical approximation of a complex model generated by mapping from a small amount of random inputs to the corresponding model outputs. Over the past few years, a number of surrogate models have been developed in the field of uncertainty quantification, for example polynomial regression model [1], radial basis function [2], Kriging [3]/Gaussian process [4], artificial neural network [5], support vector regression [6], [7], ensemble of surrogates [8] and PCE [9], [10], [11], [12], [13], [14], [15], among which PCE has received much attention for uncertainty and sensitivity analyses [9], [10], [11], [12], [13], [14], [15], [16].

First introduced by Ghanem and Spanos [17] to stochastic mechanics based on homogeneous chaos theory [18] and later generalized by Xiu and Karniadakis [19] for different types of statistical distributions (e.g. uniform, beta and gamma), the PCE approach is to represent explicitly the stochastic model response as a series of orthonormal multivariate polynomials, i.e. PC basis [17]. In this scenario, quantification of the response probability density function is equivalent to estimation of the PC coefficients that are the coordinates of the stochastic response in the basis and can be evaluated at a set of sampling points in the input space. To build a PCE, two main approaches are typically adopted: intrusive and non-intrusive. The intrusive approach requires modifying the solving scheme of the deterministic governing equations of the model [20], whereas the non-intrusive approaches such as the projection method [21] and the regression method [9], [22] compute the PC coefficients by performing repeated simulations on limited number of input–output samples. Nevertheless, for either intrusive or non-intrusive approaches, the number of model evaluations (i.e. the computational cost) required for computing the PC coefficients increases dramatically with the number of input variables and the order of expanded polynomials.

To circumvent the issue of ‘curse of dimensionality’, several efficient non-intrusive approaches have been developed in recent years, such as adaptive methods [10], [23], [24], [25] for sequentially selecting the significant basis polynomials from the full PCE, multi-fidelity methods [4], [26] for achieving accurate predictions of quantities of interest using combination of ‘low-fidelity’ and ‘high-fidelity’ PCE simulations, sparse grid based methods [27], [28] for utilizing sparse grid interpolation techniques to reduce the number of collocation points in constructing the PCE, l1-minimization methods [26], [29], [30], [31] for scanning important bases by minimizing the l1 norm of PCE coefficients while preserving the fitting accuracy, and Bayesian methods [32], [33] for generating the sparse PCE with statistical model selection criteria. These approaches have been demonstrated with considerable computational gains compared to the classical PCE method.

In the field of signal processing and data analysis, coordinate descent (CD) algorithms have received much attention due to its simplicity and efficiency in sparsity exploitation [34], [35], [36], [37]. On the other hand, the least absolute shrinkage and selection operator (LASSO) is widely adopted to perform continuous model selection and enforce sparse solutions for problems where the number of predictors exceeds the number of cases [38]. It was found by Wu and Lange [35] that both cyclic CD, also known as pathwise coordinate descent (PCD) [34], and GCD were superior to the LAR [39] in terms of efficiency, robustness and model selection for LASSO-penalized l2 regression, while GCD was substantially faster than cyclic CD for LASSO-penalized l1 regression. GCD was further developed by Li and Osher [40] with combination of Bregman iteration [41] to solve the compressed sensing problems [42], [43], [44]. Recently, Zhou et al. [45] developed an adaptive method based on partial least squares and distance correlation for building sparse PCE and compared their algorithm with LASSO-based sparse PCE, in which PCD [36] was employed.

Motivated by the preceding analysis, the present paper aims at efficiently building sparse PCEs in the context of LASSO-based regression by taking advantage of both GCD and Bregman iteration. To the best of the authors’ knowledge, there is still no research on the capability of GCD algorithms for the purpose of sparse PCE construction. It has been shown that GCD may converge to a sparse solution significantly faster than cyclic or randomized CD [35], [46], especially for high-dimensional problems. In particular, GCD applied to problems of the LASSO form can sometimes approach to an optimal solution before executing even a single pass of all coordinates [46]. This suggests that GCD has an inherent screening ability for sparse optimization, which strongly motivates its combination with sparse PCE metamodeling for tackling high-dimensional problems. The novelty and contribution of the proposed method lie in the following aspects: (1) This study is probably the first work to develop a GCD algorithm for sparse PCE metamodeling in the context of LASSO-based regression. By updating the coordinate with the largest energy decrease [40], the GCD method is straightforward, efficient and robust in sparsity exploitation. (2) The regularization parameter λ in the LASSO-based regression is commonly selected by cross-validation [35], which often leads to inefficient computation. In addition, the inefficiency is created by the dilemma that bigger λ is preferred for more accuracy whereas smaller λ gives rise to faster convergence in the coordinate updating of GCD [40]. To tackle this issue, this study incorporates Bregman iteration into the GCD to form a Bregman-iterative GCD (BGCD) algorithm structure, which not only settles the above inefficiency by adopting a moderate λ with a relatively wide range of appropriate values, but also considerably enhances the convergence and accuracy of the GCD in solving the PC coefficients. (3) Existing non-intrusive algorithms for building sparse PCEs are mainly for problems with more cases than the number of predictors. However, the proposed algorithm is devised to handle problems with far fewer cases than the number of predictors, which is of great potential to the uncertainty quantification in practical engineering.

In this paper, the sparse PCE is employed for the GSA that aims at quantifying the respective effects of different inputs and their interactions on an assigned output response. GSA can provide complete information about the model behavior when the inputs vary in the entire domain. The priority level and ranking of the inputs resulting from the GSA can be very helpful for designers in narrowing the uncertain scope of model response. Among the developed works on sensitivity analysis, Sobol’ indices have attracted greater portion of attention due to the fact that they can provide accurate information for most models [16], [47], [48], [49]. The remainder of the paper is organized as follows. Section 2 gives a brief introduction of the Sobol’ decomposition and corresponding sensitivity indices for the GSA. In Section 3, the polynomial chaos approximation of a multidimensional model is recalled. The detailed procedure on how to determine the Sobol’ indices from the PC coefficients is given in particular. Then, in Section 4, a BGCD algorithm is proposed for building the optimal sparse PCE from a given sample set. The performance of the proposed algorithm is assessed on several benchmark examples for GSA in Section 5, and the results are compared with the well-established OMP algorithm, LAR technique and two adaptive methods. Section 6 summarizes concluding remarks.

Section snippets

Sobol’ decomposition

In this section, the Sobol’ decomposition and corresponding sensitivity indices for the GSA are briefly recalled.

Let us consider a square integrable function y=Fx having n-dimensional independent inputs defined in the unit hypercube [0, 1]n. The Sobol’ decomposition of Fx into summands of increasing dimensions can be represented as follows [47], [48], [49]:Fx1,,xn=F0+i=1nFixi+1i<jnFi,jxi,xj++F1,2,,nx1,,xn,where F0 is a constant and the integral of each summand Fi1,,isxi1,,xis over any

Full PCE

The classic PCE of the model response y=Fx can be represented as follows:y=Fx=αnβαψαx,where α=α1,,αn (with αi0) is an n-dimensional index, and βα’s are unknown deterministic PC coefficients. The multivariate polynomial ψα is the tensor product of normalized univariate orthogonal polynomials:ψαx=i=1nψαiixi.

Different types of univariate orthogonal polynomials commonly used for constructing PC are listed in Table 1 [19].

In practice, the PCE in Eq. (7) is usually truncated for computational

Bregman-iterative GCD for sparse PCE

To build a sparse PCE by solving the above LASSO-based optimization problem raises two concerns: one is to develop an effective algorithm for minimizing the objective function Eβ, and the other is how to determine the regularization parameter λ. In this section, a novel algorithm based on the integration of GCD and Bregman iteration is developed for sparse PCE construction. First, GCD is newly developed to solve the LASSO-based regression problem (13) for building sparse PCE. Second, Bregman

Numerical examples

This section is devoted to the validation and evaluation of the proposed algorithm for building sparse PCE. Three benchmark functions are first considered: the Ishigami function, Sobol’ function and Morris function. The overall performances of the proposed BGCD algorithm are compared with the PCD, GCD, OMP, LAR and two adaptive algorithms [45], [10]. Both OMP and LAR are with a MATLAB implementation called SparseLab available at http://sparselab.stanford.edu/. The quality of the sparse PCE

Conclusions

In this paper, a novel algorithm based on the integration of GCD and Bregman iteration is proposed for building sparse PCE of model response. It is then used to compute Sobol’ indices for GSA. Taking advantage of the efficiency of GCD in sparsity exploitation and capability of Bregman iteration in accuracy enhancement, the proposed BGCD efficiently and accurately selects the significant basis polynomials by solving a LASSO-based regression problem (also known as general l1-minimization problem)

CRediT authorship contribution statement

Jian Zhang: Supervision, Conceptualization, Methodology, Writing - original draft, Writing - review & editing. Xinxin Yue: Methodology, Investigation, Software, Data curation, Writing - review & editing. Jiajia Qiu: Investigation, Writing - review & editing. Lijun Zhuo: Writing - review & editing. Jianguo Zhu: Writing - review & editing.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgements

This work is supported by the National Natural Science Foundation of China (Grant Nos. 11872190, 11972014), the Six Talent Peaks Project in Jiangsu Province (Grant No. 2017-KTHY-010), and the Research Start-up Foundation for Jinshan Distinguished Professorship at Jiangsu University (Grant No. 4111480003).

References (59)

  • W. Zhao et al.

    Global sensitivity analysis with a hierarchical sparse metamodeling method

    Mech. Syst. Sig. Process.

    (2019)
  • H.P. Wan et al.

    Arbitrary polynomial chaos expansion method for uncertainty quantification and global sensitivity analysis in structural dynamics

    Mech. Syst. Sig. Process.

    (2020)
  • H.G. Matthies et al.

    Galerkin methods for linear and nonlinear elliptic stochastic partial differential equations

    Comput. Methods Appl. Mech. Eng.

    (2005)
  • O.P. Le Maitre et al.

    A stochastic projection method for fluid flow: II. Random process

    J. Comput. Phys.

    (2002)
  • G. Blatman et al.

    An adaptive algorithm to build up sparse polynomial chaos expansions for stochastic finite element analysis

    Probab. Eng. Mech.

    (2010)
  • G. Blatman et al.

    Adaptive sparse polynomial chaos expansion based on least angle regression

    J. Comput. Phys.

    (2011)
  • S. Abraham et al.

    A robust and efficient stepwise regression method for building sparse polynomial chaos expansions

    J. Comput. Phys.

    (2017)
  • S. Salehi et al.

    An efficient multifidelity l1-minimization method for sparse polynomial chaos

    Comput. Methods Appl. Mech. Eng.

    (2018)
  • G.T. Buzzard

    Global sensitivity analysis using sparse grid interpolation and polynomial chaos

    Reliab. Eng. Syst. Saf.

    (2012)
  • P.G. Constantine et al.

    Sparse pseudospectral approximation method

    Comput. Methods Appl. Mech. Eng.

    (2012)
  • J. Peng et al.

    A weighted l1-minimization approach for sparse polynomial chaos expansions

    J. Comput. Phys.

    (2014)
  • K. Cheng et al.

    Sparse polynomial chaos expansion based on D-MORPH regression

    Appl. Math. Comput.

    (2018)
  • Y.C. Zhou et al.

    An expanded sparse Bayesian learning method for polynomial chaos expansion

    Mech. Syst. Sig. Process.

    (2019)
  • A. Saltelli et al.

    Variance based sensitivity analysis of model output. Design and estimator for the total sensitivity index

    Comp. Phys. Commun.

    (2010)
  • T. Homma et al.

    Importance measures in global sensitivity analysis of noninear models

    Reliab. Eng. Syst. Saf.

    (1996)
  • R. Baptista et al.

    Some greedy algorithms for sparse polynomial chaos expansions

    J. Comput. Phys.

    (2019)
  • L.M. Bregman

    The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming

    USSR Comput. Mathem. Mathem. Phys.

    (1967)
  • J. Xu et al.

    A cubature collocation based sparse polynomial chaos expansion for efficient structural reliability analysis

    Struct. Saf.

    (2018)
  • W. Chen et al.

    Analytical variance-based global sensitivity analysis in simulation-based design under uncertainty

    J. Mech. Des.

    (2005)
  • Cited by (21)

    • Efficient reliability analysis using prediction-oriented active sparse polynomial chaos expansion

      2022, Reliability Engineering and System Safety
      Citation Excerpt :

      Therefore, the main computational burden still lies in the construction of PCE using BGCD from a finite number of input-output samples, where a set of deterministic model evaluations need to be carried out. Since the original BGCD [27] employs uniformly space-filling samples for constructing PCE, a large number of model evaluations are still required in practical applications with high-dimensional random inputs, which is obviously infeasible for a computationally demanding model (e.g. finite element model). To this end, it is important for the reliability analysis to reduce the number of expensive model evaluations as much as possible using active learning strategies.

    • Blind-Kriging based natural frequency modeling of industrial Robot

      2022, Precision Engineering
      Citation Excerpt :

      The method is also known as the behavioral model or surrogate model (or response surface model, metamodel), a low computational expense mathematical model [12]. In our work, the surrogate model is developed by the high-fidelity model using, for example, Kriging (known as Gauss Process, GP) [13,14], polynomial chaos expansion (PCE) [15,16], radial basis function (RBF) [17,18] and support vector regression (SVR) [19,20]. As a combination of mathematical and statistical techniques, surrogate models have been broadly studied in engineering issues [21].

    View all citing articles on Scopus
    View full text