1. Introduction: Problem Setting
It is well known that environmental security is a very important topic for modern society. The development of reliable and sustainable mathematical models has a significant role at this area. Identifying the most influential factors, such as chemical reaction rates, boundary conditions, and emission levels, through sensitivity analysis techniques provides invaluable insights for enhancing the model’s performance. This process not only aids in pinpointing areas requiring more detailed scrutiny but also guides improvements in the overall model structure and parametrization. By understanding which factors have the greatest impact on model outputs, researchers can prioritize efforts to refine these critical components, leading to more accurate and reliable predictions. This will lead to increased reliability and robustness of predictions obtained from large-scale environmental and climate models.
The input data for sensitivity analyses were obtained during runs of a large-scale mathematical model for the remote transport of air pollutants. The
Unified
Danish
Eulerian
Model (UNI-DEM) was developed at the Danish National Environmental Research Institute (
http://www2.dmu.dk/AtmosphericEnvironment/DEM/, 1 April 2015) [
1].
A mathematical model depicts an initial problem, frequently expressed through partial differential equations (PDEs) or analogous mathematical structures. After establishment, this model undergoes discretization via methods like finite elements, finite differences, or boundary elements. Significantly, post-discretization, some extensive mathematical models can comprise billions or even trillions of algebraic equations. Resolving these models can require multiple days, even with the use of high-performance computing systems. Hence, the creation of efficient algorithms becomes essential.
Comprehensive models depict both critical and less significant phenomena concerning the primary output of the model. Despite equal treatment, it can transpire that addressing a minor phenomenon demands substantial computational resources yet contributes marginally to the solution. An illustration of this situation is provided in this study.
Moreover, the model’s output might exhibit high sensitivity to a singular measurable input parameter. Possessing prior knowledge regarding the importance of certain parameters can prove immensely advantageous. For example, understanding that a particular parameter substantially influences the model’s result implies that it should be measured with increased accuracy. This might necessitate resource allocation for acquiring more precise measuring devices, consequently elevating the overall quality and dependability of the model’s forecasts.
Let us assume the mathematical model at hand can be described using a function
which is the vector of input parameters with a joint
probability
density
function (PDF)
. Let us assume that the input variables are independent (non-correlated) and the density function
is known, even if
are not actually random variables. This implies that the output
is also a random variable as it is a function of the random vector
with its own PDF. The above presentation is quite general. One must consider that, in most cases of large-scale modeling, the function
is not explicitly available. Frequently, the function
represents a solution to a system of partial differential equations with specified boundary and initial conditions. These initial conditions are crucial for ensuring the existence of a unique solution to the system. When dealing with large-scale problems involving multiple parameters in differential equations, demonstrating the existence of a unique solution becomes more difficult. In such cases, it is common to assume the existence of a unique solution and proceed by discretizing the problem using a suitable numerical method.
Global sensitivity analysis (GSA) offers significant benefits for both model developers and users, enabling them to quantify uncertainties in model outputs, assess the relative contributions of various input factors to these uncertainties, and prioritize efforts for their reduction. There are several critical stages involved in conducting a reliable and efficient sensitivity analysis. One of the first steps is metamodeling or approximation, which serves as a vital connection between generating experimental data and applying mathematical techniques for sensitivity analysis. Accurate approximation of data is essential for the overall reliability of the resulting sensitivity indices; therefore, identifying an effective method for approximating discrete functions is crucial. Detailed information on the approximation stage of the mathematical model under consideration can be found in [
2].
The subsequent important step in performing sensitivity analysis involves selecting an appropriate technique (e.g., local approaches like one-at-a-time experiments, screening methods, variance-based methods, or derivative-based global sensitivity approaches [
3]). This generally requires robust sampling techniques, such as Latin hypercube sampling, importance sampling, stratified sampling, and low-discrepancy sequences.
A strong contender for reliably analyzing models with nonlinearities is the variance-based approach [
4]. Its core principle is to estimate how variations in an input parameter or group of inputs contribute to the variance of the model output. To measure this, the total sensitivity index (TSI) is employed (see
Section 1.5). We consider this measure more suitable and reliable than others when multi-component analysis is required. Mathematically, it is defined as a multidimensional integral:
where
is a square-integrable function in
and
is a probability density function such that
. The advantages and reasons to choose a variance-based approach for SA in the current work are provided in
Section 1.4. A detailed description of the Sobol’ approach (a well-known and efficient variance-based approach for SA) is presented in
Section 1.5. Following Equation (
2), one can conclude that the last crucial step of providing reliable SA is to choose an efficient Monte Carlo (stochastic) algorithm for multidimensional integration to compute sensitivity indices of inputs in a reliable way.
From an environmental security perspective, it is crucial to examine the impact of variations in chemical reaction rates and emission levels on the results generated by the UNI-DEM. Conducting such an analysis can yield valuable insights that serve as a solid foundation for making well-informed assumptions, reasonable simplifications, or identifying parameters whose accuracy needs enhancement. This is because the model’s outputs can be highly sensitive to fluctuations in these parameters.
The primary objective of our research is to enhance the reliability of the model’s outputs by employing efficient Monte Carlo algorithms (MCAs) that are optimal for a specific category of integrands. In certain scenarios, an issue termed “loss of accuracy” arises when attempting to compute small sensitivity indices. This study introduces and evaluates specialized Monte Carlo techniques designed to mitigate these challenges, thereby improving the overall accuracy and dependability of the model’s results.
The main tasks of the present work are to
Study the sensitivity of the concentration levels of key pollutants (e.g., ) due to variations in chemical rates and emission levels. Apply unimprovable Monte Carlo algorithms for numerical integration to perform Sobol’ variance-based sensitivity analysis.
Compare the developed Monte Carlo algorithms based on Sobol’ quasi-random points [
2] with four existing approaches for multidimensional integration:
- –
Sobol’ approach carried out by a Gray code implementation and sets of direction numbers proposed by Joe and Kuo [
5];
- –
Owen’s scrambling [
6] taken from the collection of NAG C Library [
7];
- –
eFAST approach carried out via SIMLAB [
8];
- –
Plain Monte Carlo algorithm [
9].
Demonstrate the superior efficiency of the algorithms proposed by the authors in the current study;
Provide practical insights into the case study at hand;
Offer operational guidelines for estimating relatively small Sobol’ indices in presence of computational difficulties.
1.1. Computational Complexity of Algorithms
The selection of the most efficient numerical solution algorithm for solving large-scale problems is indeed crucial. Efficiency in this context typically refers to an algorithm’s ability to achieve a desired level of accuracy while minimizing computational complexity. Computational complexity is often measured by the number of arithmetic operations required to reach a specified accuracy level.
Here are some key points to consider when choosing the most efficient algorithm:
Accuracy: The algorithm should be able to provide solutions within a predefined acceptable error margin. This could be a fixed error for deterministic algorithms or a probabilistic error bound for stochastic algorithms.
Computational Complexity: Algorithms can be classified based on their time and space complexity. Time complexity measures how the execution time grows as the input size increases, while space complexity measures how much memory is used. Lower complexities generally indicate better efficiency.
Deterministic vs. Stochastic Algorithms: Deterministic algorithms produce the same output for the same input every time they run. Stochastic algorithms introduce randomness and might produce different outputs for the same input but can offer advantages like faster convergence or simpler implementations.
Arithmetic Operations: Counting the number of arithmetic operations provides a good estimate of an algorithm’s computational complexity. Fewer arithmetic operations usually mean less computation time and lower resource usage.
Memory Usage: In addition to arithmetic operations, the amount of memory required by an algorithm can also impact its efficiency. Some algorithms require significant amounts of memory, which can limit their applicability regarding systems with limited resources.
Communication Costs: For distributed computing environments, communication costs between nodes can significantly affect overall performance. Efficient algorithms minimize these costs.
Hybrid Approaches: Sometimes, combining multiple algorithms or using hybrid methods can lead to improved efficiency. For example, using a fast but approximate method initially followed by a slower but more accurate method can balance speed and precision.
In summary, the most efficient numerical solution algorithm is the one that balances accuracy, computational complexity, memory usage, and other relevant factors to best suit the specific requirements of the problem at hand. It is important to note that there is no universally “best” algorithm; the optimal choice depends heavily on the specifics of the problem being solved and the available computational resources.
One could consider comparing two classes of algorithms: deterministic algorithms and randomized (Monte Carlo) algorithms. Let I be the desired value of the integral. Assume that, for a given random variable , the mathematical expectation satisfies . Suppose that the mean value of n values of : is considered as a Monte Carlo approximation to the solution: where correspond to values (realizations) of an RV . Generally, a randomized algorithm may produce results with a given probability of error. When working with randomized algorithms, one must acknowledge that the computational result will be correct only with a certain (although high) probability. In most practical computations, it is acceptable to tolerate an error estimate with a probability less than 1.
Consider the following integration problem:
where
and
is an integrable function on
. The computational problem can be considered as a mapping of function
to
:
where
and
. We denote by
S the solution operator. The elements of
constitute the input data for which the problem must be solved; specifically, for any
.
In certain instances, there is interest in cases where the integrand
f possesses higher regularity. This frequently occurs in practical computations, where
f tends to be smooth and have bounded high-order derivatives. Under these circumstances, it is advantageous to exploit this smoothness. To accomplish this, we introduce the functional class
, defined as
Here, denotes the domain of interest, and refers to the Sobolev space of functions on with up to k-th order weak derivatives. The notation symbolizes the norm associated with the Sobolev space . By defining in this fashion, we can effectively capture the smoothness characteristics of the function f, thereby potentially enhancing the precision and computational efficiency of numerical methods designed for such smooth integrands.
Definition 1. Let d and k be integers, . We consider the class (sometimes abbreviated to ) of real functions f defined over the unit cube , possessing all the partial derivatives which are continuous when and bounded in sup norm when The seminorm on is defined as We keep the seminorm
in the notation for the functional class
since it is important for our further consideration. We call a
quadrature formula any expression of the form
which approximates the value of the integral
. The real numbers
are called weights and the
d dimensional points
are called nodes. It is clear that, for fixed weights
and nodes
, the quadrature formula
may be used to define an algorithm with an integration error
. We call a
randomized quadrature formula any formula of the following kind:
where
and
are random weights and nodes, respectively. The algorithm
belongs to the randomized class (Monte Carlo) denoted by
.
Definition 2. Given a randomized (Monte Carlo) integration formula for the functions from the space , we define the integration error by the probability error in the sense that is the least possible real number, such that and the mean square error We assume that it suffices to obtain an -approximation to the solution with a probability . If we allow equality, i.e., in Definition 2, then can be used as an accuracy measure for both randomized and deterministic algorithms. In such a way, it is consistent to consider a wider class of algorithms that encompass both classes: deterministic and randomized algorithms.
Definition 3. Consider the set of algorithms A: that solve a given problem with an integration error .
In such a setting, it is correct to compare randomized algorithms with algorithms based on low-discrepancy sequences like Sobol’ -sequences.
1.2. Sobol’ Sequences
-sequences are
uniformly distributed sequences (UDSs) The term
UDS was introduced by Hermann Weyl in 1916 [
10].
For practical applications, it is essential to identify a uniformly distributed sequence (UDS) that satisfies three key requirements [
11,
12]:
- 1.
Best Asymptotic Behavior: As n approaches infinity, the sequence should exhibit optimal asymptotic behavior.
- 2.
Well-Distributed Points: For smaller values of n, the points should be evenly distributed across the domain.
- 3.
Computational Efficiency: The algorithm used to generate the sequence should be computationally inexpensive.
By ensuring that these criteria are met, we can guarantee the effectiveness and practicality of the chosen UDSs in various computational tasks.
All
-sequences given in [
12] satisfy the first requirement. Suitable distributions such as
-sequences are also called
-nets and
-sequences in base
. To introduce them, define first an elementary
s-interval in base
b as a subset of
of the form
where
are integers and
for all
. Given two integers
, a
-net in base
b is a sequence
of
points of
such that
for any elementary interval
E in base
b of hypervolume
. Given a non-negative integer
t, a
-sequence in base
b is an infinite sequence of points
such that, for all integers
, the sequence
is a
-net in base
b.
I. M. Sobol’ [
11] defines his
-meshes and
-sequences, which are
-nets and
-sequences in base 2, respectively. The terms
-nets and
-sequences in base
b (also called Niederreiter sequences) were introduced in 1988 by H. Niederreiter [
13].
To generate the
j-th component of the points in a Sobol’ sequence, we need to choose a primitive polynomial of some degree
over the Galois field of two elements GF(2)
where the coefficients
are either 0 or 1. GF(2) is the unique field with two elements
where the addition is defined equivalently to the logical XOR operation and the multiplication to the logical AND operation, respectively. A sequence of positive integers
are defined by the recurrence relation
where ⊕ is the bit-by-bit
exclusive-or operator. The values
can be chosen freely provided that each
is odd and less than
. Therefore, it is possible to construct different Sobol’ sequences for the fixed dimension
s. In practice, these numbers must be chosen very carefully to obtain very efficient Sobol’ sequence generators [
14]. The so-called direction numbers
are defined by
. Then, the
j-th component of the
i-th point in a Sobol’ sequence is given by
where
is the
k-th binary digit of
.
Subroutines to compute these points can be found in [
15,
16]. The work [
17] contains more details.
1.3. Randomized Quasi-Monte Carlo (RQMC)
Instead of employing randomized (Monte Carlo) algorithms for computing the mentioned sensitivity parameters, one can consider deterministic quasi-Monte Carlo algorithms or randomized quasi-Monte Carlo methods [
18,
19]. Randomized (Monte Carlo) algorithms have proven highly efficient in solving multidimensional integrals over composite domains [
9,
11]. Simultaneously, quasi-Monte Carlo (QMC) methods based on well-distributed Sobol’ sequences present a compelling alternative to Monte Carlo algorithms, particularly for smooth integrands and relatively low effective dimensions (up to
) [
20,
21,
22].
Sobol’
-sequences are strong contenders for efficient QMC algorithms. Although these algorithms are deterministic, they emulate the pseudo-random sequences used in Monte Carlo integration. A significant limitation of
-sequences is their potential for suboptimal two-dimensional projections, implying that the distribution of points can deviate substantially from uniformity. If the computational problem involves such projections, the lack of uniformity could result in a considerable loss of accuracy. To mitigate this issue, randomized QMC methods can be employed. Several randomization techniques exist, with scrambling being a notable example. The primary motivation behind scrambling [
6,
23] was to enhance the uniformity of quasi-random sequences in high dimensions, as assessed through two-dimensional projections. Moreover, scrambling offers a simple and unified way to generate quasi-random numbers for parallel, distributed, and grid-based computational environments. Essentially, scrambled algorithms can be regarded as Monte Carlo algorithms with a specific choice in the density function. Thus, it is logical to compare two classes of algorithms:
deterministic and
randomized.
Various versions of scrambling methods exist, based on digital permutations, and their differences lie in the definitions of the permutation functions. Examples include Owen’s nested scrambling [
6,
24], Tezuka’s generalized Faure sequences [
25], and Matousek’s linear scrambling [
26]. Following the introduction of Niederreiter sequences [
13], Owen [
6] and Tezuka [
25] independently developed two influential scrambling methods for
-sequences in 1994. Owen specifically highlighted that scrambling can be employed to generate error estimates for quasi-Monte Carlo (QMC) methods. Numerous other techniques for scrambling
-sequences have since been proposed, many of which represent modifications or simplifications of the original Owen and Tezuka schemes. Owen’s method is particularly effective for
-sequences, whereas the Tezuka algorithm was shown to be efficient for
-sequences. Most existing scrambling methods involve randomizing a single digit at a time. In contrast, the approach presented in [
27] randomizes multiple digits within a single point simultaneously, offering enhanced efficiency when utilizing standard pseudo-random number generators as scramblers.
Owen’s scrambling [
6], also known as nested scrambling, was developed to provide a practical error estimate for QMC by treating each scrambled sequence as a separate and independent random sample from a family of randomly scrambled quasi-random numbers. Let
be quasi-random numbers in
, and let
be the scrambled version of the point
. Suppose that each
can be represented in base
b as
, with
K being the number of digits to be scrambled. Then, nested scrambling proposed by Owen [
6,
24] can be defined as follows:
and
, with independent permutations
for
, where
means the permutation used for
. The permutation used for the next element depends on the value of
, and it is denoted by
. The permutation applied for
for
depends on all the previous values, and it is denoted by
. Of course,
-net remains
-net under nested scrambling. However, nested scrambling requires
permutations to scramble the
l-th digit.
The rate for scrambled-net Monte Carlo is
in probability, while the rate for unscrambled nets is
or
along
sequences [
28]. The first rate is an average case result for a fixed function
f taken over random permutations. Other findings pertain to the worst-case performance over functions given a fixed set of integration points. Since scrambled nets retain their net structure, these worst-case bounds are also applicable to them [
28]. Some scrambling methods do not alter the asymptotic discrepancy of quasi-random sequences [
6]. Despite improving the quality of quasi-random sequences, this enhancement is not immediately evident in the computation of the
discrepancy. Moreover, it is presently impossible to theoretically demonstrate that one scrambled quasi-random sequence outperforms another. While scrambling does not impact the theoretical bounds on the discrepancy of these sequences, it does enhance the measures of two-dimensional projections and the evaluation of high-dimensional integrals.
At its core, Owen’s nested scrambling [
6] relies on the randomization of a single digit during each iteration, and it is a powerful technique applicable to all
-sequences. However, from an implementation perspective, nested scrambling—or what is often referred to as path-dependent permutations—requires significant bookkeeping and can lead to more complex implementations. On the other hand, it has been established that its convergence rate is
. While this rate is commendable, it is still not optimal, even when dealing with integrands in
. Removing the logarithmic term from the estimate would make the rate optimal. Nevertheless, it remains an open question whether this estimate is precise, i.e., whether the logarithm can indeed be eliminated. Importantly, the demonstrated convergence rate for Owen’s nested scrambling algorithm surpasses that of unscrambled nets, which stands at
. Consequently, it becomes essential to conduct numerical comparisons between our algorithms and Owen’s nested scrambling.
1.4. Concept of Sensitivity Analysis
The process of sensitivity analysis (SA) is of crucial importance for large-scale mathematical models. This process includes the following three steps:
first, one should define the probability distributions for the input parameters under consideration;
second, samples should be generated according to the defined probability distributions using a proper sampling strategy;
third, an efficient approach is necessary for sensitivity analysis to be applied to study the output variance according to the variation in the inputs.
Moreover, there are additional stages during the process of providing sensitivity analysis to a particular mathematical model: (i) approximation, which is an important link between the generation of experimental data and the mathematical technology for sensitivity analysis, and (ii) use of a proper probability approach for computing specific sensitivity measures.
A variety of sensitivity analysis techniques are documented [
29]. Most current methods for conducting SA hinge on particular assumptions about the model’s behavior (such as linearity, monotonicity, and additivity of the relationship between model inputs and outputs). These assumptions commonly apply to a wide array of mathematical models. Nonetheless, some models incorporate substantial nonlinearities and/or stiffness, thereby invalidating assumptions about linearity and additivity. This issue is particularly pronounced when working with nonlinear systems of partial differential equations. Our research focuses on the UNI-DEM, a mathematical model that simulates the transport of air pollutants and other substances across extensive geographical regions. The system of partial differential equations captures key physical phenomena like advection, diffusion, deposition, and both chemical and photochemical processes affecting the studied species. Additionally, emissions and fluctuating meteorological conditions are considered. Nonlinearity in the equations predominantly stems from modeling chemical reactions [
1]. If the model outputs are sensitive to a particular process, it suggests that the process may require refinement or more accurate modeling. Our aim is to enhance the reliability of the model predictions and identify processes needing further scrutiny as well as input parameters that demand higher-precision measurements. Thorough sensitivity analysis is crucial for uncovering potential areas where model simplification can occur. As a result, the development and exploration of more precise and robust sensitivity analysis methodologies are essential.
Among quantitative methods, variance-based techniques are frequently employed [
4]. A literature review based on high-impact-factor journals from Science and Nature (Thomson Reuters, Journal citation reports, April 2015) and all Elsevier journals using Scopus bibliometric search tools for publications from 2005 to 2014 is presented in [
30]. This study yields the following conclusions:
There is a progressively increasing proportion of global sensitivity analysis approaches, although local techniques remain predominant.
Regression- and variance-based techniques are the most favored.
Medicine and chemistry are the leading scientific domains in applying global sensitivity analysis.
Two prominent variance-based methods were utilized: the Sobol’ approach and the extended Fourier Amplitude Sensitivity Test (eFAST). These methods were implemented using Monte Carlo algorithms (MCAs) or the SIMLAB software tool for sensitivity analysis [
8]. In the eFAST method developed to estimate the total effects, the variance of the model output
y (a
d-dimensional integral) is expressed as a single-dimensional integral with respect to a scalar variable
s. This transformation involves representing each input variable
as
, where
represents a suitable set of transformations and
are integer frequencies (refer to [
31] for detailed information).
Sobol’ indices offer advantages over eFAST, particularly in computing higher interaction terms in an analogous way to main effects. In the Sobol’ measure, each effect (main or interaction) is determined by evaluating a multidimensional integral using the Monte Carlo method. The strength of the Sobol’ method lies in its capacity to calculate the total sensitivity index with only one Monte Carlo integral per factor.
The assessment of parameter importance can be investigated through numerical integration in the context of variance analysis (ANOVA). Several unbiased Monte Carlo estimators for global sensitivity indices were devised, leveraging ANOVA decomposition of the model function [
32,
33,
34,
35].
1.5. Sobol’ Approach
In variance-based sensitivity analyses, the primary indicator associated with a given normalized input parameter
(normalized) is typically defined as
where
represents the variance of the conditional expectation of
given
and
denotes the total variance of
. This indicator is termed the
first-order sensitivity index by Sobol’ [
36] or
correlation ratio by McKay [
37]. The
total sensitivity index [
38] quantifies the overall effect of a given parameter, encompassing all possible coupling terms between that parameter and all others.
The
total
sensitivity
index (TSI) of an input parameter
is defined in the following way [
36,
38,
39]:
where
is called
the main effect (first-order sensitivity index) of
and
is the
order sensitivity index. Higher-order terms capture the interaction effects between the unknown input parameters
, where
on the output variance. Later on, we show how sensitivity indices
are defined via the variances of conditional expectations
.
The global SA method employed in this work is grounded in decomposing an integrable model function
f in the
d-dimensional factor space into terms of escalating dimensionality [
36]:
where
is a constant. The total number of summands in Equation (
6) is
. The representation Equation (
6) is referred to as the ANOVA representation of the model function
if each term is chosen to satisfy the following condition [
33]:
An important comment here is that, if the entire presentation Equation (
6) of the right-hand side is used, then it does not simplify the initial problem. The main expectation is that a truncated sequence
where
, is considered as a sufficiently good approximation to the model function
f. Then,
The quantities represent the total and partial variances, respectively. They are obtained by squaring and integrating over
Equation (
6), assuming that
is a square-integrable function (ensuring that all terms in Equation (
6) are also square-integrable functions). Consequently, the total variance of the model output is partitioned into partial variances in a manner analogous to the model function, resulting in a unique ANOVA decomposition:
The use of probability theory concepts is based on the assumption that the input parameters are random variables distributed in
, which defines
, and also as random variables with variances regarding Equation (
7). For example,
is presented by a conditional expectation:
Based on the above assumptions about the model function and the output variance, the following quantities
are referred to as the global sensitivity indices [
33]. For
, this formula coincides with Equation (
4), and the so-defined measures correspond to the main effect of input parameters as well as the effect of the cross-interactions.
Based on Formulas (
7) and (
8), it becomes apparent that addressing the problem of global sensitivity analysis mathematically involves calculating total sensitivity indices (Equation (
5)) of the relevant order. This process requires computing multidimensional integrals of the form in Equation (
2). In general, obtaining
entails computing
integrals of type in Equation (
7).
Earlier discussions highlighted that the fundamental assumption behind representation Equation (
6) is that the essential features of model functions (
1), which describe typical real-world scenarios, can be captured using low-order subsets of input variables. These subsets include terms up to order
, where
. Leveraging this assumption allows us to reduce the dimensionality of the original problem. Following Sobol’ [
33], we consider any set of
m variables (
):
and let
be the set of
complementary variables. Thus,
. Let
.
The variances corresponding to the subsets
and
can be defined as
where the complement of the subset
K in the set of all parameter indices is denoted by
. The left sum in Equation (
9) is extended over all subsets
, where all indices
belong to
K. Then, the total variance, corresponding to the subset
, is
and is extended over all subsets
, where at least one
.
The procedure for computing global sensitivity indices is based on the following representation of the variance:
(see [
33]). The equality (
10) enables the construction of a Monte Carlo algorithm for evaluating
and
:
where
is a random sample and
corresponds to the input subset denoted by
.
For example, for and :
Calculating the impact of higher-order interactions can be achieved through an iterative process. For instance, consider the following equation: Here, can be determined if the corresponding first-order sensitivity indices have already been computed.
2. Monte Carlo Algorithms Based on Modified Sobol’ Sequences
An algorithm following a
shaking technique was proposed recently in [
40]. This concept involves taking a Sobol’
point
of dimension
d. The point is treated as the center of a sphere with a radius
. A random point
is selected uniformly on the sphere. A random variable
is defined as a value of the corresponding integrand at that random point, i.e.,
.
Next, the random points are chosen, where is a unique uniformly distributed vector in . The radius is relatively small , so remains within the same elementary interval , where the pattern point resides. Using a subscript i in indicates that the i-th point is within it. Thus, if , then too.
It was proven in [
40] that the mathematical expectation of the random variable
coincides with the value of the integral (
3); that is,
It allows for defining a randomized algorithm. One can take the Sobol’
point
and
shake it somewhat.
Shaking means to define random points
according to the procedure described above. For simplicity, this algorithm is abbreviated as MCA-MSS.
The probability error of the MCA-MSS was examined in [
2]. It was demonstrated that, for integrands with continuous and bounded first derivatives, specifically
, where
, the following relationship applies:
where the constants
and
do not depend on
n.
Here, a modification of the MCA-MSS is proposed and analyzed. The new algorithm will be called MCA-MSS-S.
It is assumed that
,
. The unit cube
is divided into
disjoint sub-domains, such that they coincide with the elementary
d-dimensional subintervals defined in
Section 1.2 with
for all
.
Thus, in each d-dimensional sub-domain , there is precisely one point . If the random point remains within , i.e., after shaking, one may attempt to use the smoothness of the integrand f if .
Then, if
is a probability density function, such that
, then
where
are constants. If
is the diameter of
, then
where
are other constants.
In the particular case when the subintervals are with edge for all constants, we have and . In each sub-domain , the central point is denoted by , where .
Let us say we choose two points,
and
. The point
is selected during our MCA-MSS procedure, while
is chosen to be symmetrical to
with respect to the central point
in each cube
. This approach results in a total of
random points. We can compute all function values
and
for
and approximate the value of the integral as follows:
This estimate corresponds to the MCA-MSS-S. We will demonstrate later that this algorithm exhibits an optimal rate of convergence for functions with bounded second derivatives, i.e., for functions
, whereas the MCA-MSS has an optimal rate of convergence for functions with bounded first derivatives:
.
One can prove the following:
Theorem 1. The quadrature Formula (11) constructed above for integrands f from satisfiesand where the constants and do not depend on n.
Proof of Theorem 1. For the fixed
point
, one can use the
d-dimensional Taylor formula to present the function
in
around the central point
. Since
, there exists a
d-dimensional point
lying between
and
such that
where
and
To simplify matters, the superscript
of the argument in the last two formulas is omitted, assuming that the formulas apply to the
cube
. Now, we can rewrite Formula (
12) at the previously defined random points
and
, both located within
. This yields
where
is another
d-dimensional point lying between
and
. Adding Equations (
13) and (
14), we obtain
Due to the symmetry, there is no term involving the gradient
in the previous formula. If we examine the variance
, keeping in mind that the variance of the constant
is zero, we arrive at
Given that , it is possible to enhance the last inequality by substituting the terms and with the seminorm L (while removing the front bracket) and replacing the products and with the squared diameter of the sub-domain .
Now, let us return to the notation with superscripts, bearing in mind that the preceding considerations apply to an arbitrary sub-domain
. The variance can be estimated as follows:
Therefore, the variance of
can be estimated as follows:
Thus,
Following Tchebychev’s inequality to the variance (
15) leads to the following estimation
for the probable error
, where
, which concludes the proof. □
It is noteworthy that the Monte Carlo algorithm MCA-MSS-S possesses an optimal rate of convergence for functions with continuous and bounded second derivatives [
9]. This signifies that the rate of convergence (
) cannot be enhanced for the functional class
within the class of randomized algorithms
.
Note that both the MCA-MSS and MCA-MSS-S have a single control parameter, namely the radius of the sphere of shaking. However, effectively utilizing this control parameter increases the computational complexity. The challenge arises because, after shaking, the random point might leave the multidimensional sub-domain. Consequently, after each such operation, one must verify whether the random point remains within the same sub-domain. Verifying if a random point lies within a given domain is computationally expensive when dealing with a large number of points. A minor modification to the MCA-MSS-S can help to overcome this difficulty.
Instead of shaking, one can simply generate a random point
uniformly distributed inside
, and then take the symmetric point
according to the central point
. This fully randomized approach simulates the MCA-MSS-S, but the
shaking occurs with different radii
in each sub-domain. We call this algorithm SS-MCA as it resembles the stratified symmetrized Monte Carlo method [
11]. Clearly, the SS-MCA is less computationally expensive than the MCA-MSS-S, but it lacks a control parameter like the radius
, which can be considered as a randomly chosen parameter in each sub-domain
.
It is important to note that all three algorithms, the MCA-MSS, MCA-MSS-S, and SS-MCA, have optimal rates of convergence for their respective functional classes. Specifically, the MCA-MSS is optimal in , while both the MCA-MSS-S and SS-MCA are optimal in .