In practical applications, experimental regions are often subject to multiple constraints. Linear constraints involve linear relationships between variables, while quadratic constraints involve quadratic terms. Geometrically, linear constraints form a hyperplane in higher-dimensional spaces, while quadratic constraints form a surface, making the feasible region more complex. Linear constraints are generally simpler to compute, and they often describe limitations on costs, time, or resources in practical applications. Quadratic constraints are used for more complex problems, such as when variables have non-linear relationships. For example, in portfolio optimization, the objective is to maximize returns. Using uniform designs for simulation, investment portfolios that yield higher returns can be identified. Suppose that the total investment amount is b, and there are s types of securities available for investment. Based on the prior information, the effect of each security is represented by , where for , and the corresponding weights are denoted by . A linear constraint can be formulated as . Geometrically, this constrained region corresponds to the intersection of and the given hyperplane. When additional prior information is available, multiple linear constraints can be introduced, expressed as , for . In this case, when and , the constrained region is simplified to a line segment within the unit cube . In physical engineering systems, experimental factors may exhibit quadratic relationships. In such cases, the constraint can be expressed as , which represents the intersection of a hypersphere and . When dealing with high-dimensional variables and limited resources, a uniform design is an effective approach to extracting a small number of samples to identify patterns within a specified region. A practical strategy involves first generating points that are uniformly distributed in . These points are then mapped to the target region through an appropriate transformation, ensuring that the resulting points maintain a uniform distribution within the target region.
3.1. One Arbitrary Linear Constraint
We aim to identify experimental points that follow a uniform distribution using the distribution function of random vectors with a uniform distribution. A key challenge in this process is to determine the measure of the experimental region, which involves computing high-dimensional integrals. Based on the work presented in Lasserre [
23], which utilized the Laplace transform method to compute the measure of the intersection between a simplex in
and
, we derive the following results.
Lemma 1. In the non-empty bounded space, , where for ; then, the measure of iswhere Here, denotes a vector in , where the entries at the 1st, -th, … and -th positions are equal to 1, and all other positions are equal to 0.
Proof. We begin by proving that Lemma 1 holds when
for
and
. Let
, where
and
By the existence theorem for Laplace transforms, the Laplace transform
of
exists when the parameter
s, or the real part of
s, is positive. The Laplace transform of
is defined as
:
, where
or the real part of
is positive. According to the definition of the Laplace transform, we have
By applying Fubini’s theorem, we obtain
According to the proposition for the Laplace transform,
is the Laplace transform of the function
. Therefore, we have
When
the set
is non-empty. Taking
, we have
Next, we will prove that Lemma 1 holds when
for
,
for
and
. Let
where
and
. When
, it is clear that
. When the parameter
s or the real part of
s is positive, the Laplace transforms
of
and
of
both exist. Define the Laplace transforms of
and
as
Similarly, let
; then, we have
When , the set is non-empty and the formula is well defined.
If , the terms with can be excluded first, and then Lemma 1 can be applied to obtain the measure of the corresponding experimental region. □
For convenience in subsequent calculations, we can derive the following result for the integral computation based on Lemma 1.
Lemma 2. Let ; then, we havewhere , is a vector in with 1s at the positions and 0s at all other positions. Proof. Based on Lemma 1, by representing the measure in Lemma 1 as a high-dimensional integral, we treat some of the variables in the constraint conditions as constants. Through a coordinate transformation, we can obtain Lemma 2. □
Lemmas 1 and 2 provide the foundation for determining the transformation that maps the uniformly distributed points from to a uniform distribution within the target region.
First, we consider the case where the experimental region is subject to an arbitrary single linear constraint. Define , where . Through a straightforward analysis, experimental regions with a single arbitrary equality constraint can be classified into the following cases:
The intercept term b is zero, and the coefficients consist of positive, negative, and zero values: .
The intercept term is non-zero, with some coefficients being positive while the rest are zero. Let us now examine the case where
. If all coefficients are positive, we arrange them in ascending order and denote the smallest coefficient as
. If
,
, the corresponding variable
is treated as a slack variable, allowing the experimental region to be relaxed into
However, if
,
, the corresponding variable
is treated as a slack variable, and the experimental region fully coincides with its intersection with the
:
The intercept term is non-zero, and not all coefficients are positive: , where and , with }.
For case (i), let
be a random vector uniformly distributed on
with a density function
where
represents the measure of
. By Lemma 1, we can obtain
Following the approach outlined in
Section 2, we assume that the marginal distribution function of
is
where
Additionally, the conditional distribution can be expressed as
Building on the concept from
Section 2, let each component of the uniformly distributed random vector
in the unit hypercube correspond to the distribution function and conditional distribution function of
; we can derive the transformation formula by combining (
1) and (
2):
where
are the test points uniformly distributed on the
. In Equation (3), we first assume that
is uniformly distributed in the experimental region. Then, the vector composed of its distribution function
and conditional distribution function
follows a uniform distribution in the unit hypercube of the same dimension. Thus, each component of the uniformly distributed random vector
in the unit hypercube corresponds to the distribution function and conditional distribution function of
. The components that are not subject to linear constraints can be directly represented by the corresponding components in
. By utilizing the linear constraints, we can express the slack variable
, and this relationship can be obtained in the equation.
We now turn our attention to the uniformity of the distribution of experimental points obtained through the transformation (
3). Based on the above results, we can derive the following result.
Theorem 1. In the non-empty bounded space , let be a random vector uniformly distributed over . Then, the random vector obtained through the transformation in (3) is uniformly distributed over . Proof. 1.Let
and
represent random vectors uniformly distributed over
and
, respectively, and satisfying (
3). The distribution function of
is given as
Let
represent a random vector uniformly distributed over
, respectively. We know that
is a random vector over the space
. Define
as an invertible transformation from
to
such that
, where
is the inverse of the marginal distribution function
of
, and
are the inverse of the conditional distribution functions
of
, respectively. The joint distribution functions of
and
are given by
Thus, we have
From the above discussion, we can conclude that
Therefore,
, obtained through the transformation (
3), is uniformly distributed over
. □
From Theorem 1, we can conclude that the transformed test points are uniformly distributed over .
For case (ii), consider the case where
and
. Treat
as a slack variable, and the new experimental region can be represented as
. Let
be a uniformly distributed random vector in
. We can derive the density function
and the measure of the region
Then, by Lemma 2, the corresponding marginal distribution function can be expressed as
where
and the conditional distribution functions
where
Building on the concept from
Section 2, let each component of the uniformly distributed random vector
in the unit hypercube correspond to the distribution function and conditional distribution function of
. The third equation in the expression can be obtained from the constraint condition; then, we can derive the transformation formula by combining (
4) and (
5):
where
are the test points uniformly distributed on the
. We now turn our attention to the uniformity of the distribution of the experimental points obtained by the transformation (
6). Based on the previous results, we have the following conclusion.
Theorem 2. In the non-empty bounded space with , let be a uniformly distributed random vector over . Then, the random vector obtained through the following transformation (6) is also uniformly distributed over . Thus, by Theorem 2, it is established that the transformed test points are also uniformly distributed over .
When
, we have
. In this scenario, we can treat
as a slack variable, and the new experimental region can be represented as
. Let
be a uniformly distributed random vector in
. The density function can then be expressed as
, where
. By Lemma 2, the corresponding marginal distribution function can then be expressed as
Additionally, the conditional distribution functions for
can be expressed as
Similarly, we can derive the design using the transformation
where
are the test points uniformly distributed on
. We now turn our attention to the uniformity of the distribution of the experimental points obtained by the transformation (
9).
Theorem 3. In the non-empty bounded space , let be a random vector uniformly distributed over , and suppose that . Then, the random vector obtained through the transformation (9) is also uniformly distributed over . From Theorem 3, we can conclude that the transformed test points are uniformly distributed over
. When all coefficients are equal to 1, this conclusion degenerates into the result of uniform designs over the standard simplex presented in Fang and Wang [
7].
For the third case, consider the condition
. Treat
as a slack variable, and the new experimental region can be represented as
. Let
be a uniform random variable in
. The density function
can then be obtained, and the measure of the region
is given by
By Lemma 2, the corresponding marginal distribution function can be expressed as
and the conditional distribution functions
where
Similarly, we can derive the design using the transformation
Based on the previous results, we can derive the following result.
Theorem 4. In the non-empty bounded space , let be a random vector uniformly distributed over . Then, the random vector obtained through the following transformation (12) is also uniformly distributed over . The proofs of Theorems 2–4 can be established in a manner similar to that of Theorem 1, and we omit them. Thus, by Theorem 4, we have established that the transformed test points are also uniformly distributed over .
Based on the aforementioned results, the steps for identifying a uniform design over the experimental region
, subject to an arbitrary linear constraint, can be summarized in Algorithm 1.
Algorithm 1 General construction algorithm for one arbitrary linear constraint |
- 1:
Input: The experiment domain ; - 2:
Step 1: Compare the magnitude of b with 0. If , proceed to step 2. If , divide both sides of the constraint equation by b to obtain a new constraint: , where , Then, proceed to step 3; - 3:
Step 2: If all (or for , then is empty. Otherwise, reorder such that and . Then, transform into ; - 4:
Step 3: If all for , arrange them in ascending order. If , transform into ; If , transform to ; If a rearrangement exists such that and , with , transform into ; otherwise, is empty; - 5:
Step 4: Calculate the corresponding marginal distribution function ( 1) (or ( 4), ( 7), ( 10)) and the conditional distribution functions ( 2) (or ( 5), ( 8), ( 11)) on (or , , ); - 6:
Step 5: Given a set of uniformly distributed test points in generated using existing methods (such as the good lattice method), randomly permute each point and denote it as . Then, for each point, we can obtain distinct design configurations uniformly distributed over using the transformation ( 3) (or ( 6), ( 9), ( 12)); - 7:
Output: Compare the CCD values of the designs and select the one with the smallest CCD as the final design points, denoted by .
|
In Step 2, based on the idea of slack variable models, as discussed by Schneider et al. [
18] and Schneider et al. [
19], we can choose
as a slack variable in a linear programming problem. Using the existing constraints, the new experimental domain can be expressed as
. Similarly, in Step 3, the new experimental domains can be denoted as
,
and
. After identifying the uniformly distributed test points within these experimental domains, it remains to prove that these design points are also uniformly distributed in the original experimental domain.
In the following, we provide an example to facilitate a better understanding of Algorithm 1.
Example 1. By Algorithm 1, we obtainand We can solve Equations (13) and (14) to obtainand Thus, based on constraint , we have Using the modified good lattice point method to generate design points in , the corresponding design points on are obtained from Equations (15)–(17). The construction results are visualized in Figure 1, which shows the results of the IRT, AR, and SR methods for constructing a 10, 50, and 100-run uniform design on . In Figure 1, the plus sign represents the design points obtained using the IRT-based method, the circle represents the design points obtained using the AR method, and the asterisk represents the design points obtained using the SR method. Based on Figure 1, we can directly observe that the design points obtained using the method proposed in this paper, based on the IRT method, are more widely and evenly distributed in the experimental space. In practical experiments, we can find the nearest design points to arrange the experiments. Moreover, we run experiments for to compare the IRT, AR, and SR methods, with numerical results presented in Table 1. We observe that the IRT method consistently outperforms the other two methods. Additionally, the AR method takes the longest time. Both the method proposed in this paper and the SR method consume relatively less computational time. However, the designs generated using the proposed method have a smaller CCD scores in the experimental region, indicating better uniformity. 3.2. Multiple Arbitrary Linear Constraints
Regarding the case where the number of distinct linear constraints t (where ) exceeds one, the feasible region for selecting experimental points is the intersection of multiple -dimensional hyperplane segments. In this case, algebraic methods can be used to compute the normal vector of the intersecting hyperplanes, thereby representing the general solution of intersecting parts. The problem is reduced to the case of a single linear constraint.
The experimental region
is considered, where
. Given the number of non-redundant linear constraints in this experimental region, we can directly determine the experimental points
within the region
, assuming that
is non-empty. Alternatively, we can express the components of
that are unconstrained by the linear conditions and then use the IRT method to construct a uniform design on
. However, due to the arbitrary values of
and
t, it is challenging to intuitively determine the number of non-redundant linear constraints within the experimental region, making it difficult to establish the dimension of
and identify which components of
are constrained. A natural approach is to view the constraints in the experimental region as a system of linear equations over the real number field, denoted as (I):
, where
A is the coefficient matrix of the system, expressed as
is the unknown vector in the linear system (I), and is the constant vector. The augmented matrix of the system is indicated by , and the dimension of the solution space of system (I) corresponds to the dimension of the experimental space . The dimension d of the experimental space can be determined based on the rank (where ) of the coefficient matrix A, the rank (where ) of the augmented matrix , the number of linear constraints t, and the total dimensionality s of the space. Specifically, if the rank of the design matrix A is smaller than the dimension s of , the dimension of the experimental region is the absolute difference between s and ; that is, .
When , system (I) becomes a homogeneous system of linear equations, where . Based on the solvability of homogeneous linear systems, if or , the system will have non-zero solutions. If , the system will have only the trivial zero solution. When , system (I) is a non-homogeneous system of linear equations. Based on the solvability of non-homogeneous systems, if , the system will have non-zero solutions. If , the system will have a unique solution. If , the system will not have a solution.
This construction process can be summarized in Algorithm 2.
Algorithm 2 General construction algorithm for multiple arbitrary linear constraints |
- 1:
Input: The experimental domain }, where ; - 2:
Step 1: Express the constraints as a system of linear equations and determine whether solutions exist for the system; - 3:
Step 2: Find the general solution to the system and use it to determine the normal vector of the hyperplane segment; - 4:
Step 3: Use the obtained normal vector to represent the experimental region with a single linear equality constraint, then apply Algorithm 1 to obtain the design; - 5:
output: The final design points .
|
In Step 1, it is essential to first analyze the magnitude of under conditions and , as well as to derive the expressions for the experimental points in the space. Subsequently, the uniformly distributed experimental points can be determined for each case accordingly.