Abstract
Optimization problems with PDE constraints are widely used in engineering and technical fields. In some practical applications, it is necessary to smooth the control variables and suppress their large fluctuations, especially at the boundary. Therefore, we propose an elliptic PDE-constrained optimization model with a control gradient penalty term. However, introducing this penalty term increases the complexity and difficulty of the problems. To solve the problems numerically, we adopt the strategy of “First discretize, then optimize”. First, the finite element method is employed to discretize the optimization problems. Then, a heterogeneous strategy is introduced to formulate the augmented Lagrangian function for the subproblems. Subsequently, we propose a three-block inexact heterogeneous alternating direction method of multipliers (three-block ihADMM). Theoretically, we provide a global convergence analysis of the three-block ihADMM algorithm and discuss the iteration complexity results. Numerical results are provided to demonstrate the efficiency of the proposed algorithm.
MSC:
49M41; 65K10
1. Introduction
Elliptic PDE-constrained optimization problems are of significant importance in many fields, including engineering, physics, biology and finance. In engineering, they are essential for optimizing designs in structural mechanics, fluid dynamics, and electromagnetics. In physics, they are crucial to modeling heat transfer, quantum mechanics, and elasticity. In biology, these problems are integral in utilizing modeling diffusion processes, population dynamics, and biochemical reactions. Furthermore, in finance, they are vital for option pricing, risk management, and portfolio optimization. The versatility and applicability of elliptic PDE-constrained optimization problems make them a cornerstone for solving complex real-world challenges in many disciplines. For more details, see [1,2,3,4]. The general model is as follows:
where , . The domain is convex, open, and bounded. The boundary is - or polygonal. and denote the desired state and the source term, respectively. The parameters , . The operator L denotes the uniformly elliptic differential operator:
where , , , and for a specific constant , the following inequality holds:
With the development of science and engineering, PDE-constrained optimization problems in practical applications have become increasingly complex. In particular, the gradient penalty term plays a crucial role in PDE-constrained optimization problems. Referring to the total variation model in image restoration [5], Gao and Ding introduced a gradient penalty term to the objective function. This modification improves the model by ensuring the smoothness of the control forces while still minimizing the total control forces [6]. Moreover, Clever and Lang proposed the idea of adding a gradient penalty term to the state temperature, augmenting the objective function with a function that depends on the state gradient. This approach aims to minimize thermal stress within a glass [7]. The physical significance of the gradient penalty term is to constrain the gradient changes of the solution and promote its smoothness and continuity. This penalty term helps prevent abrupt changes or oscillations in the solution, particularly in problems requiring stability and continuity, such as fluid dynamics or structural optimization. By penalizing large gradient changes, the gradient penalty term guides the optimization process towards solutions exhibiting desired characteristics like smoothness and stability.
Inspired by the efficiency of the gradient penalty term and in order to smooth and suppress large changes in the control variable u, we propose an elliptic PDE-constrained optimization model with a control gradient penalty term and pointwise box constraints on the control:
To numerically solve (1), we apply the “First discretize, then optimize” strategy in [2]. The first step is to discretize the continuous elliptic PDE-constrained optimization problem. Specifically, we employ the finite element method due to its high efficiency. The second step is to select an optimization algorithm. Existing optimization algorithms for solving PDE-constrained optimization problems can be divided into two categories. One comprises the higher-order methods, also known as Hessian-type methods, which mainly include the semi-smooth Newton (SSN) method [8,9], the sequential quadratic programming (SQP) method [6] and so on. The other category comprises the first-order methods, also known as gradient-type methods, which mainly include the fast iterative shrinkage thresholding (FISTA) algorithm [10], the alternating direction multiplier method (ADMM) [11,12,13] and the accelerated block coordinate descent (ABCD) method [14]. Among them, the higher-order algorithms usually have locally super-linear convergence rates and can obtain high-accuracy solutions, but they are also computationally expensive. On the other hand, first-order algorithms have gained significant attention due to their simple iterative techniques, generally linear or sublinear convergence rates, and very low computational costs. In particular, the advantages of first-order optimization algorithms are more obvious when solving large-scale non-smooth optimization problems. This motivates us to further investigate the application of first-order methods in solving PDE-constrained optimization problems.
Inspired by the effectiveness of ADMMs in solving large-scale finite-dimensional optimization problems, as well as the latest work on applying an ADMM to PDE-constrained optimization problems [11,12,13], we focus on using an ADMM. In certain situations, computing the exact solution of each subproblem may be either impossible or extremely costly, even when it is theoretically possible. To overcome this difficulty, inexact ADMM algorithms have been extensively studied. In [15], Eckstein et al. first proposed an inexact version of an ADMM. After that, a proximal ADMM (PADMM) was proposed to make the subproblems easier to solve in [16]. For an inexact version of the PADMM algorithm, see [17,18]. Moreover, in [19], Chen et al. presented the sGS-imsPADMM method, which combines the inexact two-block semi-proximal ADMM with the inexact symmetric Gauss–Seidel (sGS) method. Then, Song et al. introduced an inexact heterogeneous ADMM (ihADMM) algorithm for sparse PDE-constrained optimization problems with -control cost in [11]. Unlike a standard ADMM, ihADMM defines the augmented Lagrangian function for the two subproblems using different weighted inner products. This algorithm was later applied to elliptic optimal control problems with -control cost, as described in [20]. Later, in [12], Chen et al. developed a multilevel ADMM (mADMM) algorithm by incorporating a multilevel approach with inexact subproblem solutions, which substantially lowered computational overhead. Following the success of this multilevel approach, a multilevel heterogeneous ADMM (mhADMM) was further proposed in [13] to tackle optimal control problems with -control cost.
Motivated by the efficiency of ADMM-type algorithms, extending the classical two-block ADMM to the multiblock case is a natural idea. However, as shown in [21], multiblock extensions generally do not guarantee convergence, with counterexamples illustrating nonconvergence. Despite this, there are specific problem structures where multiblock ADMMs can still achieve convergence. In this paper, we propose a three-block inexact heterogeneous ADMM (three-block ihADMM) algorithm for solving a new elliptic PDE-constrained optimization model with a control gradient penalty term in the objective function. This model introduces a control gradient penalty term in the objective function and pointwise box constraints on the control. By incorporating these features, unlike the simple extension of existing methods, our approach is specifically designed to address the complex structure of this problem. The three-block ihADMM employs a heterogeneous strategy to formulate the augmented Lagrangian function and uses Krylov-based methods to inexactly solve the saddle point problem associated with the u-subproblem, and the other two subproblems have closed-form solutions, enhancing both the flexibility and applicability of the algorithm in handling complex optimization challenges. Theoretically, we rigorously prove the convergence of the three-block ihADMM algorithm and establish an iteration complexity of , demonstrating that our approach offers a solution with solid convergence guarantees. This contribution provides both computational efficiency and theoretical rigor for solving multiblock optimization problems, particularly in cases where convergence has traditionally been challenging.
This paper is organized as follows. In Section 2, a new three-block ihADMM algorithm is proposed. Specifically, we first employ the standard piecewise linear finite element to discretize the original continuous problem. Then, we combine the heterogeneous strategy and the inexact strategy to propose the use of the three-block ihADMM algorithm for the discretized problem. In Section 3, we provide the theoretical convergence analysis and discuss the iteration complexity. Numerical experiments are presented in Section 4. In Section 5, we summarize the paper.
2. A Three-Block Inexact Heterogeneous Alternating Direction Method of Multipliers (Three-Block ihADMM)
2.1. An Inexact ADMM Algorithm in Continuous Space
To apply ADMM-type methods, we introduce artificial variables v and z to rewrite (1) in the following equivalent form:
From Proposition 1 in [13], we know that the elliptic PDE involved in (1) has a unique weak solution. Let denote the continuous linear solution operator; then, the solution can be expressed as . The adjoint operator : is also continuous and linear. It is evident that (1) exhibits strong convexity. Hence, from the equivalence between (1) and (2), it is obvious that the existence of and uniqueness of the solution to (2) are guaranteed. The optimal solution, denoted as , can be characterized by the first-order optimality condition as follows.
Theorem 1.
denotes the optimal solution of (2) if and only if there exists an adjoint state, , and Lagrange multiplier, , ensuring that the following conditions are satisfied in the weak sense:
By employing the solution operator , we can reformulate (2) into the following reduced form:
where and . The indicator function is defined as
Consequently, the augmented Lagrangian function for (3) is as follows:
where represent Lagrangian multipliers and are penalty parameters. The classical two-block ADMM in Hilbert space can be directly extended to the three-block version. Specifically, given the initial point , the ADMM iterative scheme is as follows:
In general, computing an exact solution for each subproblem may be expensive and not always necessary. Thus, employing iterative methods like Krylov-based techniques such as the conjugate gradient (CG) method [22] and the generalized minimal residual (GMRES) method [23] to obtain approximate solutions for these subproblems is an effective strategy. These methods are particularly effective for solving large, sparse linear systems, as they avoid the high computational costs associated with direct methods like Gaussian elimination. These iterative methods yield approximate solutions within a specified tolerance, which is beneficial when exact solutions are either unnecessary or too costly to compute. Furthermore, Krylov-based methods can be enhanced through preconditioning techniques that improve the systems’ condition number, making them highly flexible and effective for various problems. Building on this concept, we propose the three-block inexact ADMM algorithm in continuous Hilbert space for (3), as detailed in Algorithm 1.
Notice that the u-subproblem of Algorithm 1 is a convex differentiable optimization problem. By omitting the error vector , it can be equivalently reformulated as the following system:
Notice that , let , so we can obtain
We point out that (4) can be rewritten as
Furthermore, we can eliminate p to obtain the reduced system:
where I denotes the identity operator. It is evident that the linear system (5) can be viewed as a particular instance of the generalized saddle point problem. Given the structure of this linear system, various Krylov-based methods can be utilized to inexactly solve it.
| Algorithm 1 Three-block inexact ADMM algorithm for (3) |
Input: Choose the initial point and the parameter Let the sequence , satisfying Set Output: Step 1: Compute an approximation solution, , of
such that the residual
satisfies Step 2: Compute Step 3: Compute Step 4: Compute Step 5: Compute Step 6: If the stopping criteria are met, stop. Otherwise, set and return to Step 1. |
Moreover, it is obvious that there exists a closed-form solution to the z-subproblem:
Although we have given Algorithm 1 in function space, we need to discretize (1) to implement the numerical computation. The fact that both the u-subproblem and the z-subproblem of the ihADMM algorithm in function space are well structured provides some insight into performing numerical discretization. The v-subproblem also has a special structure, which we will discuss in the following subsection. In other words, we have to utilize the structure (1) in function space when performing numerical discretization. Then, extending the well-formed structure, like (5) and (6), in parallel to the discretized problem, thus solving the discretized problem with an efficient ADMM-type algorithm, is an important starting point for the next subsection of this paper.
2.2. Finite Element Discretization
To numerically solve (1), we discretize both y and u using the standard piecewise linear finite element method [24]. Next, we begin by considering a regular and quasi-uniform triangulation family, , of . For each , we denote the diameter of the set T by . Additionally, we define as the diameter of the largest ball that fits entirely within Let denote the mesh size. We assume that the triangulation satisfies the standard regularity conditions typically used in finite element error estimates.
Assumption 1
([20], Assumption 2). There exist two positive constants, and , such that , . In addition, let , , and denote its interior and its boundary, respectively. In the case that Ω is a convex polyhedral domain, we have In the case that Ω is a domain with a - boundary Γ, we assume that is convex and that all boundary vertices of are contained in Γ, such that , where denotes the measure of the set and is a constant.
Given the homogeneous boundary condition of the state equation, we define the discretized state space
and the discretized control space
where represents the space of polynomials of degree less than or equal to 1.
For a given with nodes denote the nodal basis functions, which span and and satisfy
Then, we have
Let
represent the relative coefficient vectors. Define
as the nodal projection of and onto , where and . Let
denote the relative coefficient vectors. In addition, let the discretized feasible set be defined as
and define the bilinear form:
Then, we can formulate the discretized version of (1):
From the error estimation results in [25,26], we arrive at the following result.
Theorem 2.
Let represent the optimal control solution of (1) and represent the optimal control solution of (7). Then, we have
Proof.
The proof is similar to Proposition 4.5 in [26], so we omit it here. □
To rewrite (7) in its matrix vector form, we define the following stiffness, mass and lump mass matrices:
Moreover, we have
where
Based on the matrix form mentioned above, it is evident that and are symmetric positive definite matrices, and is a symmetric matrix.
Consequently, based on the above representation, we can rewrite (7) in the following matrix vector form:
Similar to (1), we know (8) is strictly solvable. Moreover, we can derive the important discrete first-order optimality conditions.
Theorem 3.
denotes the optimal solution of (8) if and only if there exist Lagrangian multipliers, and , such that the following conditions hold:
For the convenience of representation, let
Then, the problem (8) can be rewritten as
The augmented Lagrangian function is defined as
Then, the iteration format of the ADMM is as follows:
Notice that the -subproblem is equivalent to the following linear system:
Let
represent the discretized state and
represent the discretized adjoint state, respectively. Consequently, (12) can be reformulated as
By eliminating the variable , (13) can be reformulated as follows:
Notice that (14) contains the inverse of the matrix , which inevitably leads to additional computations. To reduce the computational cost, we introduce a heterogeneous strategy. Specifically, we employ various weighted inner products to formulate the augmented Lagrangian functions. On one hand, for the -subproblem and -subproblem, given the penalty parameters , we define the corresponding augmented Lagrangian function based on an -weighted inner product as follows:
On the other hand, for the -subproblem, we define the corresponding augmented Lagrangian function based on the -weighted inner product as follows:
Based on the above augmented Lagrangian functions, we write the iterative format of the ADMM:
where denotes the step size.
Benefiting from such a heterogeneous strategy, each subproblem of the three-block ihADMM algorithm can be efficiently implemented. Next, we provide the specific iteration process.
For the -subproblem, let
then, we have
Consequently, the solution of the -subproblem can be given as follows:
Moreover, it is easy to see that the -subproblem has a closed-form solution:
In addition, notice that the -subproblem at the kth iteration is equivalent to the linear system:
Similar to the steps in (12) and (13), (16) can be rewritten as
By eliminating the variable , (17) can be reformulated in a reduced form without incurring any additional computational cost:
We point out that (18) is a special case of the generalized saddle point problem. Therefore, Krylov-based methods can be employed to inexactly solve the problem. The GMRES method [23] is a highly efficient technique for solving linear systems. The benefits of using GMRES to solve saddle point problems include its ability to efficiently handle large and sparse linear systems, its convergence properties, and its capability to handle nonsymmetric and indefinite matrices. GMRES can also be flexible in terms of preconditioning and can be parallelized to improve computational efficiency. Let denote the residual error vector satisfying
In the numerical experiments, we require that the residual vector satisfies
to guarantee that the error vector satisfies
Moreover, to improve the convergence rate of the GMRES method, we employ the PMHSS preconditioner in [27]. Specifically, the PMHSS preconditioner is defined as
where .
We point out that the -subproblem involves solving a large-scale linear system, which is the primary computational expense of the algorithm. The -weighting technique can approximate a system with blocks to one with blocks without any extra computational cost. The -weighting inner product also provides a closed-form solution for the -subproblem. Additionally, the -weighting inner product gives the -subproblem a separable structure, allowing for a closed-form solution that can be computed with a single projection operator.
In Algorithm 2, we present the matrix vector form of the three-block inexact heterogeneous ADMM (three-block ihADMM) for (11).
| Algorithm 2 Three-block inexact heterogeneous ADMM (three-block ihADMM) for () |
Input: Choose the initial point and the parameter Let the sequence satisfying . Set Output: Step 1: (18) by the GMRES method. Step 2: Step 3: Step 4: Step 5: Step 6: If the stopping criteria are met, stop. Otherwise, set and return to Step 1. |
Furthermore, we notice that each subproblem of Algorithm 2 is a discretized form of each subproblem of Algorithm 1, which illustrates the benefit of using different weighted inner products for discretization.
3. Convergence Analysis
We will show the convergence analysis and discuss the iteration complexity of Algorithm 2 in this section. For computational convenience, we introduce the following two lemmas.
Lemma 1
([20], Appendix). There are two basic identities:
where and are from the same Euclidean space, and Q denotes a self-adjoint positive semidefinite linear operator.
Lemma 2
([19], Lemma 6.1). If the sequence satisfies
Then, we have
To facilitate the following iteration complexity analysis, we define the function
:
We know from the definitions in (10) that , , and are proper, convex, and closed functions. Since the matrices and are both symmetric and positive definite, the gradient operators and are strongly monotone. Therefore, we have
where matrices and are also symmetric and positive definite. In addition, we know that is a maximal monotone operator, i.e.,
Let
denote the exact solutions at the th iteration in (15). Then, the gap between and is shown in the following lemma.
Lemma 3.
Let be the sequence generated by Algorithm 2 and be the sequence defined by (23). Then, for any , we have
where , , and the error tolerance .
Proof.
According to the optimality conditions, we have
respectively.
Therefore,
Then, (24a) holds. The projection is nonexpansive; thus, we have
Furthermore,
Therefore, we complete the proof. □
Subsequently, we define
Furthermore, we introduce the following two key propositions that are crucial for analyzing the global convergence and discussing the iterative complexity of the three-block ihADMM algorithm.
Proposition 1.
Let denote the sequence generated by Algorithm 2 and denote the Karush–Kuhn–Tucker (KKT) point of problem (11). Then, for any , we have
Proof.
We know from the optimality conditions for that
respectively. Thus, using Theorem 3, (26a)–(26c) and taking , , , , and in (21a), (21b), and (22), we obtain
Adding the above three inequalities together, we obtain
where we have used the fact that
Then, we will reformulate the last four terms on the left side of (27). First, using (19a), we obtain
Second, we can obtain the following result by employing (19b) and :
Then, substituting (28), (29), and (30) into (27), we can obtain (25). □
Proposition 2.
Let denote the sequence generated by Algorithm 2 and denote the KKT point of (11). are defined by (23). Then, for any , we have
Proof.
Similar to Proposition 1, we can finish the proof by replacing , , and with , , and in the proof of Proposition 1. □
Theorem 4.
Let denote the KKT point of (8) and denote the sequence generated by Algorithm 2. Let be the associated state and be the adjoint state. Then, for any and and , it follows that
Furthermore, there is a constant C that depends only on and , such that
where is defined in (20).
Proof.
We first demonstrate the global convergence of the iteration sequences. First, we need to prove that the sequence is a bounded sequence. To achieve this, we define the sequences and as follows:
For any and , we know from Proposition 2 that . Consequently, we obtain
Using Lemma 3, we can obtain
Then we have
Hence, for any , we have
Given that , it follows that , ∀. Then, the sequences and are both bounded sequences. Given the definition of and , it follows that the sequences , , , and are also bounded. Additionally, the boundedness of the sequence is ensured by the updating rules for and . Therefore, the sequence is bounded. This implies that there exists a subsequence converging to an accumulation point . Next, we will demonstrate that satisfies the KKT conditions, which are actually equal to .
By applying Proposition 2, we derive that
which means
We know from Lemma 3 that
By taking the limits of both sides of (37a)–(37c), we can derive from that
Obviously, according to (38), we know that , , and are convergence sequences, and their limits are ,, and , respectively. Therefore, we have , , and . Thus, all we need to do next is to show that In order to achieve this, we take the limits for in (26a) to obtain
This leads to
Consequently, we know from (9a) that
Similarly, we have
which results in
Then, from (9b), we know that ; thus,
Moreover, to finalize the proof, we have to demonstrate that the sequence converges to . According to (36), it follows that
Because , we have
Then, we obtain
Therefore, we have established the convergence of , thus completing the proof of (32). For the proof of (33), it follows straightforwardly from the definition of , and we omit the detailed steps here.
Finally, we give the proofs (34) and (35), which provide the iteration complexity results for the sequence generated by the three-block ihADMM algorithm. First, we know from the optimality conditions (26a)–(26c) for that
We denote Then, according to the definition of , we have
where
Next, we search for an upper bound of . First, based on the definition of and (36), we can easily obtain for that
Then, we can obtain an upper bound of :
Then, we know from (31) in Proposition 2 that
Hence, we have
Then, by substituting (40) into (39), we obtain
Thus, by applying Lemma 2, (34)–(35) hold. As a result, by integrating the global convergence results, we have finalized the proof for Theorem 4. □
4. Numerical Experiments
In this section, we demonstrate the numerical results of incorporating a control gradient penalty term into the cost function of elliptic PDE-constrained optimization problems. We utilize the proposed three-block ihADMM algorithm to solve the problems. For comparison, the numerical results derived from the classical ADMM algorithm are also provided.
4.1. Algorithmic Details
First, we outline the algorithmic details that are common to both examples.
Discretization: The standard piecewise linear finite element method was employed for the discretization process.
Initialization: We chose zero as the initial value.
Parameter setting: The step length was chosen as for both algorithms. For comparison, in each example, the two algorithms used the same parameters, and . The parameter was selected as the optimal parameter based on extensive numerical experiments. When using the three-block ihADMM algorithm to solve elliptic PDE-constrained optimization problems with a control gradient penalty term, the choice of the control gradient penalty coefficient is crucial for several reasons:
1. Numerical stability: Too large a value for the parameter can lead to numerical instability. This may manifest as slow convergence or even failure to converge, as well as erratic behavior in the numerical solution.
2. Convergence speed: If is too large, it may slow down the convergence rate of the algorithm, requiring more iterations to meet the convergence criteria.
3. Impact on solution quality: The choice of the penalty term coefficient affects the quality of the final optimal solution. An excessively large coefficient may overly penalize the solution, causing it to deviate from the true optimal solution of the problem. Conversely, a coefficient that is too small may result in suboptimal precision in the optimization outcome.
Therefore, it is essential to adjust and balance the coefficient of the control gradient penalty term appropriately to ensure stable, fast convergence to an accurate optimal solution.
Terminal condition: Let h denote the mesh size. In our numerical experiments, cases with different values for h were considered. Let #dofs represent the number of degrees of freedom associated with the control variables in each grid layer. To evaluate the accuracy of an approximate optimal solution, we assessed the corresponding KKT residual error. The algorithms were terminated when , where
and
The maximum iteration number was set to 600.
Computational environment: The results presented in this paper were generated using MATLAB R2023b with the FEM package iFEM [28] running on a desktop computer with Intel Core i7 CPU (2.10 GHz) with 32 GB of RAM.
4.2. Examples
Example 1.
(Example 3.3 in [2] with an additional control gradient penalty term)
Consider
where . The desired state and the parameters are set to and
In this example, the penalty parameter , , and were chosen as , , and . The exact solutions of this problem cannot be determined in advance. Therefore, we utilized numerical solutions obtained on a grid with as the benchmark reference.
The numerical optimal controls obtained by the three-block ADMM algorithm on the grid with are displayed in Figure 1.
Figure 1.
The numerical optimal control, , with different values of parameter for Example 1. (a) with . (b) with . (c) with . (d) with .
From Figure 1, we can see that when , the numerical optimal control, , is not smooth in the cross-section. Meanwhile, when , the cross-section becomes smooth. Moreover, as the parameter of the gradient penalty term increases, the numerical optimal control, , becomes smoother and smoother. This shows that adding a control gradient penalty term to the PDE-constrained optimization problem results in a solution with smoothing properties. Moreover, we would like to point out that a larger penalty term will strongly discourage large control gradients, leading to a smoother control solution, while a smaller penalty term may lead to undesirable smoothness in the optimization results. However, the magnitude of the control gradient penalty term should be chosen carefully to balance the desired smoothness of the control strategy with the specific requirements and constraints of the problem.
As an example, we show the numerical results of in Table 1. For the proposed three-block ihADMM algorithm and the classical ADMM algorithm, the mesh size, h; the number of degrees of freedom, #dofs; the KKT residual, ; the CPU time; and the number of iterations, #iter, are reported. We would like to point out that each case in Table 1 represents a different size of mesh. In each case, the three-block ihADMM algorithm and the classical ADMM algorithm were terminated with the same stopping criteria. Therefore, #iter and CPU time were compared when the KKT residuals for each mesh were of the same order of magnitude.
Table 1.
The convergence behavior of the three-block ihADMM algorithm and the classical ADMM algorithm.
From each row of Table 1, it is evident that the three-block ihADMM algorithm requires fewer iterations and is significantly faster than the classical ADMM algorithm in obtaining approximate solutions across different final mesh sizes. Specifically, the fifth and sixth columns of Table 1 show that the three-block ihADMM algorithm outperforms the classical ADMM algorithm in terms of both CPU time and the number of iterations as the final mesh size decreases. Furthermore, the results show that as the discrete mesh size becomes finer, the number of iterations and computational time required by the classical ADMM algorithm increase faster than for the three-block ihADMM algorithm. This highlights the efficiency of the proposed three-block ihADMM algorithm in achieving moderate accuracy while reducing computational cost.
Example 2.
(Example 4.1 in [9] with an additional control gradient penalty term)
Consider
where . The parameters are set as and . The desired state was chosen as , where . S represents the solution operator associated with .
The penalty parameter , , and were chosen as , , and . The numerical optimal control on the grid with is displayed in Figure 2.
Figure 2.
The numerical optimal control, , with different values of parameter for Example 2. (a) with . (b) with . (c) with . (d) with .
We can see from Figure 2 that when , the numerical optimal control, , is not smooth in the cross-section at points a and b. When , the upper boundary and lower boundary become smooth. Specifically, as the parameter of the gradient penalty term increases, the numerical optimal control, , becomes smoother and smoother at the upper and lower boundaries. We can see that introducing a control gradient penalty term into the objective function of PDE-constrained optimization problems can result in a smoother control curve.
As an example, we present the numerical results for for both the three-block ihADMM algorithm and the classical ADMM algorithm in Table 2.
Table 2.
The convergence behavior of the three-block ihADMM algorithm and the classical ADMM algorithm.
The results presented in Table 2 indicate that the proposed three-block ihADMM algorithm demonstrates superior numerical efficiency regarding both the number of iterations and computational time when compared to the classical ADMM algorithm. Furthermore, the fifth and sixth columns of Table 2 show that as the mesh size decreases, the three-block ihADMM algorithm exhibits a more significant advantage in terms of the number of iterations and the computational time compared to the classical ADMM algorithm.
5. Conclusions
In this paper, a new elliptic PDE-constrained optimization model with a control gradient penalty term and pointwise box constraints on the control is proposed to address the application requirements. The main purpose of adding the control gradient penalty term is to smooth the control variable, particularly at the boundary. Nevertheless, adding this penalty term makes the system more complex to solve. To tackle this problem, we introduce two artificial variables and propose a convergent three-block ihADMM algorithm. Unlike the classical ADMM algorithm, the proposed algorithm introduces the inexactness strategy. Moreover, different weighted inner products are adopted to define the augmented Lagrangian function in three subproblems. Theoretically, we present the global convergence analysis and provide theoretical results on the iteration complexity for the proposed algorithm. The numerical experiments demonstrate that the three-block ihADMM algorithm significantly outperforms the classical ADMM algorithm in terms of both the number of iterations and computational time across various final mesh sizes. Moreover, as the final mesh size decreases, the efficiency of the three-block ihADMM algorithm becomes even more obvious.
Author Contributions
Conceptualization, X.C. and X.S.; methodology, X.C. and X.S.; software, T.W.; validation, X.C.; formal analysis, T.W. and X.C.; writing—original draft preparation, X.C. and T.W.; writing—review and editing, X.S.; visualization, T.W.; supervision, X.S.; funding acquisition, X.C. and X.S. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the National Key R&D Program of China (No. 2023YFA1011303), the National Natural Science Foundation of China (No. 12301477, No. 12301479, No. 42274166), the Fundamental Scientific Research Projects of Higher Education Institutions of Liaoning Provincial Department of Education (No. JYTMS20230165), and the Fundamental Research Funds for the Central Universities (No. 3132024202).
Data Availability Statement
The data presented in this study are available upon request from the corresponding author.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Huang, F.; Chen, Y.; Chen, Y.; Sun, H. Stochastic collocation for optimal control problems with stochastic PDE constraints by meshless techniques. J. Math. Anal. Appl. 2024, 530, 127634. [Google Scholar] [CrossRef]
- Hinze, M.; Pinnau, R.; Ulbrich, M.; Ulbrich, S. Optimization with PDE Constraints; Springer: Berlin, Germany, 2009; pp. 171–172. [Google Scholar]
- Zheng, Q. A reordering-based preconditioner for elliptic PDE-constrained optimization problems with small Tikhonov parameters. Comp. Appl. Math. 2023, 42, 169. [Google Scholar] [CrossRef]
- Sirignano, J.; MacArt, J.; Spiliopoulos, K. PDE-constrained models with neural network terms: Optimization and global convergence. J. Comput. Phys. 2023, 481, 112016. [Google Scholar] [CrossRef]
- Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Phys. D 1992, 60, 259–268. [Google Scholar] [CrossRef]
- Gao, L.; Ding, J. An improved DMOC method with gradient penalty term. Adv. Mater. Res. 2014, 945, 2784–2787. [Google Scholar] [CrossRef]
- Clever, D.; Lang, J. Optimal control of radiative heat transfer in glass cooling with restrictions on the temperature gradient. Optim. Control Appl. Methods 2012, 33, 157–175. [Google Scholar] [CrossRef]
- Hinze, M.; Vierling, M. The semi-smooth Newton method for variationally discretized control constrained elliptic optimal control problems implementation, convergence and globalization. Optim. Methods Softw. 2012, 27, 933–950. [Google Scholar] [CrossRef]
- Liu, Y.; Wen, Z.; Yin, W. A multiscale semi-smooth Newton method for optimal transport. J. Sci. Comput. 2022, 91, 39. [Google Scholar] [CrossRef]
- Liang, J.; Luo, T.; Schonlieb, C.B. Improving “fast iterative shrinkage-thresholding algorithm”: Faster, smarter, and greedier. SIAM J. Sci. Comput. 2022, 44, A1069–A1091. [Google Scholar] [CrossRef]
- Song, X.; Yu, B.; Wang, Y.; Zhang, X. An FE-inexact heterogeneous ADMM for elliptic optimal control problems with L1-control cost. J. Syst. Sci. Complex. 2018, 31, 1659–1697. [Google Scholar] [CrossRef]
- Chen, X.; Song, X.; Chen, Z.; Yu, B. A multi-level ADMM algorithm for elliptic PDE-constrained optimization problems. Comput. Appl. Math. 2020, 39, 1–31. [Google Scholar] [CrossRef]
- Chen, X.; Song, X.; Chen, Z.; Xu, L. A multilevel heterogeneous ADMM algorithm for elliptic optimal control problems with L1-control cost. Mathematics 2023, 11, 570. [Google Scholar] [CrossRef]
- Chen, Z.; Song, X.; Chen, X.; Yu, B. A warm-start FE-dABCD algorithm for elliptic optimal control problems with constraints on the control and the gradient of the state. Comput. Math. Appl. 2024, 161, 1–12. [Google Scholar] [CrossRef]
- Eckstein, J.; Bertsekas, D. On the Douglas–Rachford splitting method and the proximal point algorithm for maximal monotone operators. Math. Program. 1992, 55, 293–318. [Google Scholar] [CrossRef]
- Eckstein, J. Some saddle-function splitting methods for convex programming. Optim. Methods Softw. 1994, 4, 75–83. [Google Scholar] [CrossRef]
- Bingsheng, H.; Liao, L.; Han, D.; Yang, H. A new inexact alternating directions method for monotone variational inequalities. Math. Program. 2002, 92, 103–118. [Google Scholar]
- Ng, M.K.; Wang, F.; Yuan, X. Inexact alternating direction methods for image recovery. SIAM J. Sci. Comput. 2011, 33, 1643–1668. [Google Scholar] [CrossRef]
- Chen, L.; Sun, D.; Toh, K.C. An efficient inexact symmetric Gauss-Seidel based majorized ADMM for high-dimensional convex composite conic programming. Math. Program. 2017, 161, 237–270. [Google Scholar] [CrossRef]
- Song, X.; Yu, B. A two-phase strategy for control constrained elliptic optimal control problems. Numer. Linear Algebra Appl. 2018, 25, 21–38. [Google Scholar] [CrossRef]
- Chen, C.; He, B.; Ye, Y.; Yuan, X. The direct extension of ADMM for multi-block convex minimization problems is not necessarily convergent. Math. Program. 2016, 155, 57–79. [Google Scholar] [CrossRef]
- Lasdon, L.; Mitter, S.; Waren, A. The conjugate gradient method for optimal control problems. IEEE Trans. Autom. Control. 1967, 12, 132–138. [Google Scholar] [CrossRef]
- Saad, Y.; Schultz, M.H. GMRES: A generalized minimum residual algorithm for solving nonsymmetric linear systems. SIAM J. Sci. Stat. Comput. 1986, 7, 856–869. [Google Scholar] [CrossRef]
- Mora, D.; Rodríguez, R. A piecewise linear finite element method for the buckling and the vibration problems of thin plates. Math. Comput. 2009, 78, 1891–1917. [Google Scholar] [CrossRef]
- Casas, E. Using piecewise linear functions in the numerical approximation of semilinear elliptic control problems. Adv. Comput. Math. 2007, 26, 137–153. [Google Scholar] [CrossRef]
- Wachsmuth, G.; Wachsmuth, D. Convergence and regularization results for optimal control problems with sparsity functional. ESAIM Control Optim. Calc. Var. 2011, 17, 858–886. [Google Scholar] [CrossRef]
- Cao, S.; Wang, Z. PMHSS iteration method and preconditioners for Stokes control PDE-constrained optimization problems. Numer. Algorithms 2021, 87, 365–380. [Google Scholar] [CrossRef]
- Chen, L. iFEM: An Integrated Finite Element Methods Package in MATLAB; Technical report; University of California at Irvine: Irvine, CA, USA, 2008. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).

