Elsevier

Computers & Chemical Engineering

Volume 108, 4 January 2018, Pages 276-288
Computers & Chemical Engineering

Evaluating smart sampling for constructing multidimensional surrogate models

https://doi.org/10.1016/j.compchemeng.2017.09.016Get rights and content

Highlights

  • Extensive numerical evaluation of smart sampling algorithm (SSA) is performed using a diverse test bed of analytical functions.

  • Robustness of SSA is examined against Sobol sampling over the wide ranges of dimensions and domain sizes.

  • Numerical comparison of SSA with existing adaptive approaches is illustrated.

  • SSA is employed for three process systems engineering case studies to demonstrate its practical applicability.

Abstract

In this article, we extensively evaluate the smart sampling algorithm (SSA) developed by Garud et al. (2017a) for constructing multidimensional surrogate models. Our numerical evaluation shows that SSA outperforms Sobol sampling (QS) for polynomial and kriging surrogates on a diverse test bed of 13 functions. Furthermore, we compare the robustness of SSA against QS by evaluating them over ranges of domain dimensions and edge length/s. SSA shows consistently better performance than QS making it viable for a broad spectrum of applications. Besides this, we show that SSA performs very well compared to the existing adaptive techniques, especially for the high dimensional case. Finally, we demonstrate the practicality of SSA by employing it for three case studies. Overall, SSA is a promising approach for constructing multidimensional surrogates at significantly reduced computational cost.

Introduction

Process simulators are commonly used to model, study, and analyze complex nonlinear physicochemical systems. However, such simulations are generally computationally intensive, thus, prohibiting their repeated evaluations in a typical analysis procedure. Moreover, the custom-made process simulators are often black-box in nature. Hence, no system information is available to the users without evaluating an instance of this costly simulation. On these accounts, it is beneficial to convert such high-fidelity simulations into computationally inexpensive surrogate models that capture essential features with reasonable numerical accuracy. Surrogate modeling, also known as metamodeling or response surface model, is a technique to generate a mathematical or numerical representation of a complex system based on some sampled input-output data. In a philosophical discussion on the future of computational modeling, Kraft and Mosbach (2010) highlight the importance of approximation techniques and experimental designs (sampling techniques) in tackling complex multi-scale systems. The quality of any surrogate approximation depends on a sampling technique used to generate the input-output data and a surrogate modeling technique used to build the approximation. The literature (Shan and Wang, 2010) has several forms of surrogate models like polynomial response surface model (PRSM), high dimensional model representation (HDMR), kriging, radial basis functions (RBFs), support vector regression (SVR), artificial neural networks (ANNs), etc. Furthermore, many works (Henao and Maravelias, 2011, Henao and Maravelias, 2010, Caballero and Grossmann, 2008) have employed these techniques in the context of various physicochemical systems. Nonetheless, the current work focuses on the critical evaluation of a smart and adaptive sampling approach for multidimensional surrogate construction paradigms.

Commonly used sampling techniques employ uniform, quasi-random, or systematic distributions (Pronzato and Müller, 2012, Koehler and Owen, 1996). Examples are factorial design or grid sampling, random sampling, Latin hypercube sampling, orthogonal arrays, Hammersley points, Sobol sampling (QS), etc. A recent review by Garud et al. (2017b) classifies the literature on sampling techniques into three major categories viz. static system-free, static system-aided, and adaptive-hybrid. It discusses each of them thoroughly and identifies their advantages and disadvantages. The static techniques are often prone to the curse of dimensionality. Moreover, they can result in under/oversampling and thus, resulting in poor system approximation (Garud et al., 2017a). In order to tackle these issues, a new upcoming class of modern DoE (design of experiments) called adaptive sampling (sequential sampling) has gained attention from the research community over the past few years. Adaptive sampling approach has two vital advantages over the static ones viz. low computational expense and better approximation quality (Crombecq et al., 2011a). Typically, an adaptive sampling technique starts with a small set of sample points, and then adds points sequentially based on some user-defined criterion. Such criterion involves an objective (sometimes referred as a score) that aims to fill the domain (exploration) as well as improve the overall surrogate quality (exploitation) (Garud et al., 2017a, Crombecq et al., 2011a). We summarize various adaptive approaches from the literature and their vital characteristics like the exploration and exploitation criteria, dependence on the surrogate form, and the placement approach in Table 1. Although, we only discuss the key works from the adaptive sampling literature, Garud et al. (2017b) has dedicated an entire section for their discussion and the interested readers may refer to it for further details.

Jin et al. (2002) propose two approaches, namely the maximin scaled distance (MSD) and the cross validation (CV). The former is a modification of maximin distance based sampling that utilizes system information by assigning weights to the important variables while the latter uses CV error (Kohavi, 1995) to place new sample points. The CV approach can be viewed as a maximum sampling error approach with an additional feature of clustering constraint. Crombecq et al., 2009, Crombecq et al., 2011a propose a novel and generic score based sequential strategy involving exploration and exploitation. They use a combination of derivative-based local linear approximations and Voronoi tessellations to place new sample points. Although the LOLA-Voronoi strategy has shown some promising results, it can be computationally intensive for large N. A recent work by Eason and Cremaschi (2014) proposes an adaptive sampling strategy for ANN surrogates. Instead of generating all sample points in one shot, they choose them gradually based on some score from randomly generated sample sets. The score considers the normalized nearest neighbor distance of a potential point from the current sample points and its normalized expected variance evaluated using jackknifing (Efron, 1982). Though their selection of sample points is systematic, it is still from randomly generated points. Cozad et al., 2014, Cozad et al., 2015 propose an adaptive sampling for their surrogate modeling tool called ALAMO. They add sample points one at a time to the initial sample set. For each new sample point, they solve a derivative-free optimization problem to maximize the deviation of the surrogate from the real function. This can obviously be compute-intensive, as it requires the evaluation of the real function during optimization.

To this end, the adaptive sampling techniques in the literature can be broadly classified as either score-based or optimization-based. Although the latter strategies aim at the optimal sample placement, the literature suggests that such approaches are employed only with kriging surrogate due to its ready availability of the error estimate. Furthermore, these approaches may not be suitable for a wide range of problems as the performance of kriging may drop significantly with increasing dimensions. This can be tackled by using the surrogate techniques other than kriging. However, the literature clearly points out that surrogate (kriging)-independent approaches are score-based and lack the placement optimality. Therefore, there is a need for surrogate-independent and optimization-based adaptive sampling approach which is generic, robust, and ascertains optimal sample placement. Garud et al. (2017a) address this exact conundrum by proposing a novel adaptive sampling strategy, namely smart sampling algorithm (SSA). It uses crowding distance metric to identify the unexplored regions while departure function to identify the regions with complex behavior. These two concepts are then combined into an objective to formulate a point placement optimization problem. SSA iteratively solves this optimization problem to place new sample points. SSA has been developed and presented in our previous work (Garud et al., 2017a) along with its application to one dimensional cases. In this work, we present the critical evaluation of SSA for constructing multidimensional surrogate models.

This article is organized as follows. Section 2 gives a brief overview of SSA followed by our evaluation basis and plan in Section 3. We present the numerical results in Section 4 and Section 5 shows the practical application of SSA using three case studies from the chemical and process systems engineering field. Finally, in Section 6, we present our conclusions.

Section snippets

Overview of SSA

Herein, we present a brief overview of SSA for the sake of completeness. The readers may refer to Garud et al. (2017a) for the details on the development thought-process. Let y = f(x); f:NM for D:xLxxU describe the behavior of a unit/process/system whose experimental or computational quantification is complex and computationally expensive. Thus, we need an analytical or numerical surrogate model S(x) to replace f(x) so that y  S(x). Here onwards, we denote S(x) by S for the sake of

Evaluation basis and plan

We now present a detailed plan for the evaluation of SSA for constructing multi-dimensional surrogates. For this, we use two surrogate model types and compare the performance of SSA against a variety of commonly used sampling techniques. This evaluation is performed using a diverse test bed of analytical functions. Additionally, the robustness of SSA is analyzed for wide ranges of domain sizes and dimensions. Finally, three performance metrics are employed to assure a thorough comparison of the

Comparison with Sobol sampling

We now compare the performance of SSA with QS using the performance metrics (Eqs.(9a)–(9c)) discussed earlier. Tables 6 and 7 list the averaged performance metrics computed for SSA and QS using PRSM and kriging surrogates respectively. Clearly, SSA outperforms QS for all the test functions and across all the three metrics for both the surrogates. In the case of PRSM, SSA outperformed QS with the minimum PE¯-based improvement of 9% and the average improvement of around 34% (excluding TF1 where

Case studies

Besides the numerical comparison using analytical functions, the ultimate test of a technique is its practical applicability to real-life case studies. Thus, we employ SSA to the following three cases from the literature of chemical and process systems engineering: (i) biodiesel production process, (ii) multi-component distillation column, and (iii) carbon capture unit. We follow the same evaluation procedure as described earlier in Fig. 1 and compare its performance with QS using PRSM

Conclusions

In this article, we extensively evaluated a novel adaptive sampling approach, namely smart sampling (Garud et al., 2017a) for constructing multidimensional surrogate approximations. We draw following conclusions from our numerical investigation:

  • 1.

    SSA shows an excellent performance compared to QS for approximating a variety of test functions using polynomial and kriging surrogates.

  • 2.

    It performs more robustly compared to QS over ranges of domain dimensions and edge length/s for both the surrogates

Acknowledgement

This publication is made possible by the Singapore National Research Foundation under its Campus for Research Excellence And Technological Enterprise (CREATE) programme.

References (43)

  • J.J. Sikorski et al.

    Parameterisation of a biodiesel plant process flow sheet model

    Comput. Chem. Eng.

    (2016)
  • A. Ajdari et al.

    An adaptive exploration–exploitation algorithm for constructing metamodels in random simulation using a novel sequential experimental design

    Commun. Stat.-Simul. Comput.

    (2014)
  • F. Aluffi-Pentini et al.

    Global optimization and stochastic differential equations

    J. Optim. Theory Appl.

    (1985)
  • D. Busby et al.

    Hierarchical nonlinear approximation for experimental design and statistical data fitting

    SIAM J. Sci. Comput.

    (2007)
  • K.F. Butwell et al.

    Performance of gas purification systems utilizing DEA solutions

    Laurance Reid Gas Conditioning Conference

    (1975)
  • J.A. Caballero et al.

    An algorithm for the use of surrogate models in modular flowsheet optimization

    AIChE J.

    (2008)
  • A. Cozad et al.

    Learning surrogate models for simulation-based optimization

    AIChE J.

    (2014)
  • K. Crombecq et al.

    A novel sequential design strategy for global surrogate modeling.

    Simulation Conference (WSC), Proceedings of the 2009 Winter

    (2009)
  • K. Crombecq et al.

    A novel hybrid sequential design strategy for global surrogate modeling of computer experiments

    SIAM J. Sci. Comput.

    (2011)
  • L. Dixon et al.

    The global optimization problem: an introduction

    Towards Glob. Optim.

    (1978)
  • B. Efron

    The Jackknife, the Bootstrap and Other Resampling Plans

    (1982)
  • Cited by (16)

    • Distillation process optimization: A screening-clustering assisted kriging optimization method

      2021, Chemical Engineering Science
      Citation Excerpt :

      Surrogate models, also known as metamodels, are simple models that can be substituted for “black box” models. They are constructed based on a set of known input–output data (Garud et al., 2018). In contrast to “black box” models, surrogate models can provide a relatively accurate output response at a low computational cost.

    • Managing uncertainty in data-driven simulation-based optimization

      2020, Computers and Chemical Engineering
      Citation Excerpt :

      Although all surrogate models have parameters that need to be fitted based on observed data, the first category requires that one predefines the terms of the regression function, while the second category includes generic nonlinear universal approximators, whose structure and model parameters don't have a physical meaning and could change depending on the amount of data available. This work does not aim to perform a thorough comparison between all different types of surrogate functions, as this has been done in recent work on a variety of problems (Bhosekar and Ierapetritou, 2018; Davis et al., 2018; Garud et al., 2018, 2018). In this work, we employ two different types of surrogate models that have convergent approximating qualities: (a) Neural networks, and (b) Sparse-Grid polynomial interpolation models.

    • Propagation of Parametric Uncertainty in a Conceptually Designed Bioethanol Production Process

      2020, Computer Aided Chemical Engineering
      Citation Excerpt :

      At fifteen steps, the effect of increasing the number of uncertain parameters (from 1 to 15) was analyzed by performing Monte Carlo simulations of important process metrics (unit cost -UC- and net present value -NPV-). Sampling from the distributions were performed using Halton (HS), Sobol (SS), and Latin hypercube sampling (LHS) methods (Garud et al., 2018). With each sampling method, three sets with 25, 100, and 250 samples were generated.

    • Surrogate-based black-box optimisation via domain exploration and smart placement

      2019, Computers and Chemical Engineering
      Citation Excerpt :

      However, these make the models larger and more nonlinear. A single simulation run may take from a few minutes (e.g. a process model) (Garud et al., 2018a) to a few days (e.g. a computational fluid dynamics (CFD) model) (Tsunooka et al., 2018). Besides, the commercial simulators often used for developing and simulating these high-fidelity models are essentially black boxes.

    • A new termination criterion for sampling for surrogate model generation using partial least squares regression

      2019, Computers and Chemical Engineering
      Citation Excerpt :

      Correspondingly, new points are placed in these nonlinear regions. The smart sampling algorithm developed by Garud and co-workers is one of the adaptive sampling methods (Garud et al., 2017a, 2018). Through the application of two metrics, one for exploitation and one for exploration, they identify new optimal points.

    View all citing articles on Scopus
    View full text