Next Article in Journal
An Improved Relationship between the Solution and Its Corresponding Function in Fourth-Order Neutral Differential Equations and Its Applications
Previous Article in Journal
A Novel Link Prediction Method for Social Multiplex Networks Based on Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Generalized Inexact Newton-Landweber Iteration for Possibly Non-Smooth Inverse Problems in Banach Spaces

School of Science, Dalian Maritime University, Dalian 116026, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(7), 1706; https://doi.org/10.3390/math11071706
Submission received: 10 February 2023 / Revised: 29 March 2023 / Accepted: 29 March 2023 / Published: 3 April 2023

Abstract

:
In this paper, we consider a generalized inexact Newton-Landweber iteration to solve nonlinear ill-posed inverse problems in Banach spaces, where the forward operator might not be Gâteaux differentiable. The method is designed with non-smooth convex penalty terms, including L 1 -like and total variation-like penalty functionals, to capture special features of solutions such as sparsity and piecewise constancy. Furthermore, the inaccurate inner solver is incorporated into the minimization problem in each iteration step. Under some assumptions, based on ε -subdifferential, we establish the convergence analysis of the proposed method. Finally, some numerical simulations are provided to illustrate the effectiveness of the method for solving both smooth and non-smooth nonlinear inverse problems.

1. Introduction

We are interested in solving the following ill-posed inverse problem
F ( x ) = y ,
where F : D ( F ) X Y is a possibly non-smooth nonlinear operator from Banach space X to Banach space Y . In general, (1) may have many solutions. To pick the one with desired feature, we choose a convex function Θ : X , and determine a solution x such that
D ξ 0 Θ x , x 0 : = min x D Θ D F D ξ 0 Θ x , x 0 : F x = y ,
where x 0 D Θ and ξ 0 Θ x 0 denote the initial guesses, and D ξ 0 Θ x , x 0 denotes the Bregman distance induced by Θ . Due to the error of measurement, instead of y, only perturbed data y δ is available, satisfying
y δ y Y δ
with known noise level δ > 0 . A typical property of such equations is their ill-posedness, i.e., small perturbations of data may lead to huge deviations in solutions of (1). Therefore, regularization techniques are needed to obtain a stable solution x from y δ ; see [1,2,3] and references therein.
In [4,5,6,7,8], inexact Newton regularization methods were developed for solving nonlinear inverse problems in Banach spaces, where the forward mapping F is assumed to be smooth, i.e., continuously Fréchet differentiable. This type of method updates the current iteration x n δ by applying an iterative regularization scheme to solve approximately the local linearization of (1) at x n δ , i.e.,
F x n δ x x n δ = y δ F x n δ ,
where F x n δ is the Fréchet derivative of F at x n δ . Assuming the nth iterates ξ n δ , x n δ are available, by employing the Landweber iteration in [9] to (4), the inexact Newton-Landweber iteration in [6] generates the inner iterates { ( ξ n , k δ , x n , k δ ) } by
ξ n , k + 1 δ = ξ n , k δ + μ n , k δ F x n δ * J r Y y δ F x n δ F x n δ x n , k δ x n δ , x n , k + 1 δ = arg min x X Θ x ξ n , k + 1 δ , x X * , X
with ξ n , 0 δ = ξ n δ and x n , 0 δ = x n δ and suitable step length μ n , k δ , and J r Y ( y ) : = ( 1 r y Y r ) ( 1 < r < ) denotes the duality mapping from Y to its dual Y * . Let k n δ be the smallest integer such that
y δ F x n δ F x n δ x n , k n δ δ x n δ Y < γ y δ F x n δ Y
for some 0 < γ < 1 , then the next outer iterates are defined as
ξ n + 1 δ = ξ n , k n δ δ and x n + 1 δ = x n , k n δ δ .
It has been shown in [6] that this renders a regularization method if the outer iteration (7) is terminated by the discrepancy principle. Recently, in [7,8] the authors considered an inexact Newton regularization method employing a so-called two-point gradient method [10] as inner scheme and derived the convergence result under the discrepancy principle. When X and Y are both Hilbert spaces and Θ x = x 2 x 2 2 2 , one may refer to [11,12,13,14,15,16] for some convergence and convergence rates.
Except for the tangential cone condition on the forward operator, the convergence analysis for inexact Newton regularization methods in [4,5,6,7,8] requires the continuity of the derivative F ( x ) and the exact resolution of the minimization problem in (5). However, there are some cases where the forward operator F is not Gâteaux differentiable [17], which leads to the existing inexact Newton method impractical. Moreover, the exact solution of (5) can only be found for some special Θ ; this minimization problem in general can only be solved numerically inaccurately. Therefore, it is necessary to generalize the method (5)–(7) to cover the non-smooth forward operator case and to incorporate an inner inexact solver for the minimization problem in (5).
It has been shown in [18] that the Fréchet derivative F x can be substituted by another bounded linear operator sufficiently close to F x . In [17,19], by introducing the Bouligand subderivative, the authors considered the Bouligand–Landweber iteration and the Bouligand–Levenberg–Marquardt method in Hilbert spaces. By employing a bounded operator A F (depending only on a certain point) as the replacement of F x , an extension of the Gauss–Newton method was proposed in [20,21,22]. Inspired by the previous work, in this work, we propose a generalized inexact Newton-Landweber iteration for solving inverse problems with possibly non-smooth nonlinear operators, given by
ξ n , k + 1 δ = ξ n , k δ + μ n , k δ A F * J r Y y δ F ( x n δ ) A F x n , k δ x n δ , Θ x n , k + 1 δ ξ n , k + 1 δ , x n , k + 1 δ X * , X min x X Θ x ξ n , k + 1 δ , x X * , X + ε n , k + 1 ,
where A F is a bounded operator satisfying certain conditions (see (17)), ε n , k + 1 > 0 . The inner stopping index is chosen as k n δ : = min k ˜ n δ , k max , where k ˜ n δ is the first integer such that
y δ F x n δ A F x n , k δ x n δ Y p + σ ε n , k < γ y δ F x n δ Y p
for 0 < γ < 1 , σ > 0 and given integer k max 1 . The next iterates are then constructed by ξ n + 1 δ = ξ n , k n δ δ and x n + 1 δ = x n , k n δ δ . Please note that when ε n , k + 1 = 0 , k max = , F is continuously Fréchet differentiable and A F : = F ( x n δ ) , our method (8) reduces to the method (5) in [6]. In contrast to existing methods [4,5,6,7,8], our proposed method (8) does not require the Fréchet differentiablility of F and the exact solver of the minimization problem; therefore, our method is effective for solving not only smooth but also non-smooth and nonlinear inverse problems. Under certain conditions on F and A F , based on ε -subdifferential, we develop a detailed convergence analysis of the method in Section 3. The numerical results in Section 4 demonstrate the effectiveness of our method.
The rest of this paper is built up as follows. In Section 2, we give some preliminaries on Banach spaces and convex analysis. In Section 3, we formulate our proposed method, and further elaborate well-posedness and the regularization property of the proposed method. Finally, in Section 4, we provide some numerical results to indicate the effectiveness of the method in dealing with smooth as well as non-smooth inverse problems.

2. Preliminaries

In this section, we review some basic concepts on convex analysis and Banach spaces. More details are available in [23,24,25].
Let X be a Banach space with norm · X , and X * is named its dual space. If x X and x * X * , we denote by x * , x X * , X = x * ( x ) the duality pair. Given another Banach space Y and a bounded linear operator A from X to Y , we write A * : Y * X * as its adjoint, i.e., A * y * , x X * , X = y * , A x Y * , Y for any x X and y * Y * . Set N ( A ) = { x X : A x = 0 } as the null space of A and
N ( A ) : = { ξ X * : ξ , x = 0 for all x N ( A ) }
as the annihilator of N ( A ) . When X is reflexive, there holds N ( A ) : = R ( A * ) ¯ , where R ( A * ) ¯ is the closure of R ( A * ) . On a Banach space X , for any r ( 1 , ) , the subdifferential of convex function x x X r / r at x is given by
J r X ( x ) : = { ξ X * : ξ X * = x X r 1 and ξ , x X * , X = x X r } ,
which is called the duality mapping J r X : X 2 X * of X with gauge function t t r 1 .
In addition, the duality mapping J r X , for each 1 < r < , is single-valued and uniformly continuous on bounded sets when X is uniformly smooth, in the sense that its modulus of smoothness
ρ X ( s ) : = 1 2 sup { x ¯ + x X + x ¯ x X 2 : x ¯ X = 1 , x X s }
satisfies lim s 0 ρ X ( s ) s = 0 .
For a given convex function Θ : X ( , ] with effective domain D ( Θ ) : = { x X : Θ ( x ) < } , we call Θ proper if D ( Θ ) . For ε 0 , we define the ε -subdifferential of the function Θ at x by
ε Θ ( x ) : = { ξ X * : Θ ( x ¯ ) Θ ( x ) + ξ , x ¯ x X * , X ε for all x ¯ X } .
Any element ξ ε Θ ( x ) is a ε -subgradient of Θ at x [24]. When ε = 0 , then the ε -subdifferential of Θ corresponds to the subdifferential Θ . One can see that ε Θ ( x ) implies x D ( Θ ) . If Θ is lower semicontinuous, then for any x D ( Θ ) , ε -subdifferential ε Θ ( x ) is always non-empty for any ε > 0 , see [23] (Theorem 2.4.4).
For ξ ε Θ ( x ) with ε 0 ,
D ξ Θ ( x ¯ , x ) = Θ ( x ¯ ) Θ ( x ) ξ , x ¯ x X * , X + ε , x ¯ X ,
which is called the ε -Bregman distance, induced by Θ at x in the direction ξ . Obviously, D ξ Θ ( x ¯ , x ) 0 . When ε = 0 , the ε -Bregman distance becomes the Bregman distance [24]. A proper convex function Θ : X ( , ] is uniformly convex if there exists a strictly increasing continuous function φ : [ 0 , ) [ 0 , ) with φ ( 0 ) = 0 such that
Θ ( λ x ¯ + ( 1 λ ) x ) + c 0 λ ( 1 λ ) φ ( x ¯ x X ) λ Θ ( x ¯ ) + ( 1 λ ) Θ ( x )
for any x ¯ , x X and λ [ 0 , 1 ] . Θ is p-convex if φ ( t ) = c 0 t p for some c 0 > 0 and p 2 ; see [23] (Theorem 3.5.10).
Given a proper, lower semicontinuous, convex function Θ : X ( , ] , its Legendre-Fenchel conjugate is defined by
Θ * ( ξ ) : = sup x X { ξ , x X * , X Θ ( x ) } , ξ X * .
It is easily seen that Θ * is also proper, lower semicontinuous, and convex. If, in addition, X is reflexive, then [23] (Theorem 2.4.2)
ξ ε Θ ( x ) x ε Θ * ( ξ ) Θ ( x ) + Θ * ( ξ ) ξ , x X * , X + ε .
The following lemma gives further results of p-convex functionals; refer to [23] (Corollary 3.5.11) and [26] (Lemma 2.1 and Lemma 2.3) for more details.
Lemma 1.
Let X be a reflexive Banach space and Θ : X ( , ] be proper, lower semicontinuous and p-convex with φ ( t ) = c 0 t p for some c 0 > 0 , p 2 , 1 p + 1 p * = 1 . Then,
(i)
If x D ( Θ ) and ξ ε Θ ( x ) for some ε 0 , we have
c 0 x ¯ x X p 2 D ξ Θ ( x ¯ , x ) + 2 ε , x ¯ X .
(ii)
If x D ( Θ ) and ξ ε Θ ( x ) for some ε 0 , for η X * , there holds
η , x Θ * ( ξ ) X , X * 1 p * ( 2 c 0 ) p * 1 η X * p * + ε
and therefore
x Θ * ( ξ ) X p p 2 c 0 ε .
(iii)
For x D ( Θ ) , ξ ε Θ ( x ) and η X * , there holds
Θ * ( η ) Θ * ( ξ ) x , η ξ X , X * 1 p * ( 2 c 0 ) p * 1 η ξ X * p * .
(iv)
D ( Θ * ) = X * . Θ * is Fréchet differentiable on X * and its gradient Θ * : X * X satisfies
Θ * ( η ) Θ * ( ξ ) X η ξ X * 2 c 0 p * 1
for all η , ξ X * .

3. The Method

We consider (1), where F : D ( F ) X Y is a possibly non-smooth nonlinear operator between Banach spaces X and Y . For carrying out convergence analysis, we pose the following assumptions on the convex function Θ and the operator F and A F .
Assumption 1.
Θ is a proper, weakly lower semicontinuous and uniformly convex function with p 2 satisfying (11) for some c 0 > 0 .
Assumption 2.
( a ) There is ρ > 0 such that B 2 ρ ( x 0 ) D ( F ) , where B ρ ( x 0 ) : = { x X : x x 0 X ρ } . (1) has a solution x * in D ( Θ ) satisfying
D ξ 0 Θ ( x * , x 0 ) 1 4 c 0 ρ p .
( b ) F is weakly closed on D ( F ) , i.e., for any sequence { x n } D ( F ) with x n x X and F ( x n ) v Y , there hold x D ( F ) and F ( x ) = v .
( c ) There is a constant 0 η < 1 such that
F ( x ¯ ) F ( x ) A F ( x ¯ x ) Y η F ( x ¯ ) F ( x ) Y
for all x ¯ , x B 2 ρ ( x 0 ) D ( F ) . Moreover, there is a constant B ^ > 0 such that A F B ^ .
In Assumption 2, the condition (17) can be viewed as a transformation of the tangential cone condition widely used in nonlinear regularization methods [1,25,27]. When X is reflexive, by the weak closedness of F and the lower semi-continuity and p-convexity of Θ , one can show that x in (2) exists. Moreover, it has been shown in [28] (Lemma 3.2) that x is uniquely defined. We note that our generalized inexact Newton-Landweber iteration method (8) in each iteration step involves an inaccurate inner solver of the minimization problem
x : = arg min z X Θ ( z ) ξ , z X * , X
for any ξ X * . Concerning the inexact resolution of (8), we make the following assumption.
Assumption 3.
For any given ε 0 , there is a procedure S ε : X * X for solving (18) such that for any ξ X * , the element x : = S ε ( ξ ) satisfies
Θ ( x ) ξ , x X * , X min z X Θ ( z ) ξ , z X * , X + ε .
Moreover, for each ε 0 , the mapping S ε : X * X is continuous.
We summarize generalized the inexact Newton-Landweber iteration in Algorithm 1.
Algorithm 1 Generalized inexact Newton-Landweber iteration for noisy data
Input: Parameters η < γ < 1 , μ 0 , μ 1 > 0 , k max 1 , σ > 0 and τ > 1 ; a sequence of positive numbers { ε n , k } k 0 , n 0 satisfying n = 0 k = 0 ε n , k < ; the operator A F .
Initial guess:  x 0 X and ξ 0 Θ x 0 ;
  Set n = 0 .
  while F x n δ y δ Y > τ δ do
    (i)
Set ξ n , 0 δ : = ξ n δ , x n , 0 δ : = x n δ ; the inner iterates { ( ξ n , k δ , x n , k δ ) } are constructed by
ξ n , k + 1 δ = ξ n , k δ + μ n , k δ A F * J r Y ( s n , k δ ) , x n , k + 1 δ = S ε n , k + 1 ( ξ n , k + 1 δ ) ,
where s n , k δ = y δ F ( x n δ ) A F ( x n , k δ x n δ ) and
μ n , k δ = μ ˜ n , k δ ( s n , k δ Y p + σ ε n , k ) 1 r p
with
μ ˜ n , k δ = min μ 0 s n , k δ Y p ( r 1 ) A F * J r Y ( s n , k δ ) X * p , μ 1 .
    (ii)
Take k n δ = min { k ˜ n δ , k max } , where k max 1 is a given integer and k ˜ n δ 1 is the first integer satisfying
s n , k δ Y p + σ ε n , k < ( γ y δ F ( x n δ ) Y ) p .
Update the outer iterates by
ξ n + 1 δ = ξ n , k n δ δ and x n + 1 δ = x n , k n δ δ .
Set n : = n + 1 .
end while
We denote by n δ = n ( δ , y δ ) the outer stopping index such that the discrepancy principle is fulfilled, i.e.,
F ( x n δ δ ) y δ Y τ δ < F ( x n δ ) y δ Y , 0 n < n δ .
Output: An approximate solution x n δ δ of (1).
By definition of x n , k δ in (20), we have
Θ x n , k δ ξ n , k δ , x n , k δ X * , X Θ x ξ n , k δ , x X * , X + ε n , k , x X ,
which implies that ξ n , k δ ε n , k Θ x n , k δ . This fact will be used in the forthcoming theoretical analysis. The following lemma shows that Algorithm 1 is well-defined.
Lemma 2.
Let X be reflexive and Y be uniformly smooth. Assume that Assumption 1, 2 and 3 hold. Let { ε n , k } n 0 , k 0 be a sequence of positive numbers satisfying
n = 0 k = 0 ε n , k < 1 16 c 0 ρ p .
Let β > 1 , η < γ < 1 , μ 0 > 0 , and τ > 1 be chosen such that
c 1 : = 1 β η γ 1 + η τ γ p 1 p μ 0 2 c 0 1 p 1 > 0 ,
then, there holds:
(i)
for each 0 n < n δ , k ˜ n δ < and x n , k δ B 2 ρ ( x 0 ) for all 0 k k n δ ;
(ii)
Algorithm 1 terminates after n δ < iteration steps;
(iii)
for any solution x ^ of (1) in B 2 ρ ( x 0 ) D ( Θ ) , we have
D ξ n + 1 δ Θ ( x ^ , x n + 1 δ ) D ξ n δ Θ ( x ^ , x n δ ) 3 k = 0 k n δ 1 ε n , k
for 0 n < n δ . Here we may take ε 0 = 0 since ξ 0 Θ x 0 and ε n = ε n , 0 .
Proof. 
We first show that if x n δ B 2 ρ ( x 0 ) for some 0 n < n δ , then
D ξ n , k + 1 δ Θ ( x ^ , x n , k + 1 δ ) D ξ n , k δ Θ ( x ^ , x n , k δ ) c 1 μ n , k δ ( s n , k δ Y p + σ ε n , k ) r p + 2 ε n , k + ε n , k + 1
for 0 k < k n δ . Using the definition of ε -Bregman distance (9), we can arrive at, for 0 k < k n δ ,
D ξ n , k + 1 δ Θ ( x ^ , x n , k + 1 δ ) D ξ n , k δ Θ ( x ^ , x n , k δ ) = [ Θ ( x n , k δ ) ξ n , k δ , x n , k δ X * , X ε n , k ] + [ ξ n , k + 1 δ , x n , k + 1 δ X * , X Θ ( x n , k + 1 δ ) ] ξ n , k + 1 δ ξ n , k δ , x ^ X * , X + ε n , k + 1 .
Using (10) and the definition of Θ * , we further have
D ξ n , k + 1 δ Θ ( x ^ , x n , k + 1 δ ) D ξ n , k δ Θ ( x ^ , x n , k δ ) Θ * ( ξ n , k + 1 δ ) Θ * ( ξ n , k δ ) ξ n , k + 1 δ ξ n , k δ , x ^ X * , X + ε n , k + 1 = Θ * ( ξ n , k + 1 δ ) Θ * ( ξ n , k δ ) ξ n , k + 1 δ ξ n , k δ , Θ * ( ξ n , k δ ) X * , X + ξ n , k + 1 δ ξ n , k δ , Θ * ( ξ n , k δ ) x n , k δ X * , X + ξ n , k + 1 δ ξ n , k δ , x n , k δ x ^ X * , X + ε n , k + 1 .
Since Θ is p-convex, we may use (12), (14) and the definition of ξ n , k + 1 δ to derive that
D ξ n , k + 1 δ Θ ( x ^ , x n , k + 1 δ ) D ξ n , k δ Θ ( x ^ , x n , k δ ) 2 p * ( 2 c 0 ) p * 1 ( μ n , k δ ) p * A F * J r Y ( s n , k δ ) X * p * + μ n , k δ J r Y ( s n , k δ ) , A F ( x n , k δ x ^ ) Y * , Y + ε n , k + ε n , k + 1 .
Please note that A F ( x n , k δ x ^ ) = s n , k δ + [ y δ F ( x n δ ) A F ( x ^ x n δ ) ] , in view of (3), Assumption 2(c) and the property of J r Y , we further have
D ξ n , k + 1 δ Θ ( x ^ , x n , k + 1 δ ) D ξ n , k δ Θ ( x ^ , x n , k δ ) 2 p * ( 2 c 0 ) p * 1 ( μ n , k δ ) p * A F * J r Y ( s n , k δ ) X * p * μ n , k δ s n , k δ Y r + μ n , k δ s n , k δ Y r 1 ( ( 1 + η ) δ + η y δ F ( x n δ ) Y ) + ε n , k + ε n , k + 1 .
By the definition of μ n , k δ , it follows that
( μ n , k δ ) p * A F * J r Y ( s n , k δ ) X * p * = μ n , k δ ( μ n , k δ ) p * 1 A F * J r Y ( s n , k δ ) X * p * ( μ 0 ) p * 1 μ n , k δ s n , k δ Y p * ( r 1 ) ( s n , k δ Y p + σ ε n , k ) ( p r ) ( p * 1 ) p ( μ 0 ) p * 1 μ n , k δ ( s n , k δ Y p + σ ε n , k ) r p .
Using again the definition of μ n , k δ , together with (22) and (23), we derive that
μ n , k δ s n , k δ Y r 1 ( ( 1 + η ) δ + η y δ F ( x n δ ) Y ) 1 + η τ μ n , k δ s n , k δ Y r 1 y δ F ( x n δ ) Y + η μ n , k δ s n , k δ Y r 1 y δ F ( x n δ ) Y 1 + η γ τ μ n , k δ s n , k δ Y r 1 ( s n , k δ Y p + σ ε n , k ) 1 p + η γ μ n , k δ s n , k δ Y r 1 ( s n , k δ Y p + σ ε n , k ) 1 p ( 1 + η γ τ + η γ ) μ n , k δ ( s n , k δ Y p + σ ε n , k ) r p .
Analogous to the proof of [26] (Lemma 3.1), we can show that
μ n , k δ s n , k δ Y r 1 β μ n , k δ ( s n , k δ Y p + σ ε n , k ) r p κ μ ˜ n , k δ σ ε n , k ,
where σ > 0 and κ μ 1 σ 1 with
κ = 1 , i f p r , ( β p r p 1 ) p r p , i f p < r .
By inserting the above inequality, (28), (29) into (27) and using μ ˜ n , k δ μ 1 , we obtain the estimate
D ξ n , k + 1 δ Θ ( x ^ , x n , k + 1 δ ) D ξ n , k δ Θ ( x ^ , x n , k δ ) c 1 μ n , k δ ( s n , k δ Y p + σ ε n , k ) r p + κ μ ˜ n , k δ σ ε n , k + ε n , k + ε n , k + 1 c 1 μ n , k δ ( s n , k δ Y p + σ ε n , k ) r p + 2 ε n , k + ε n , k + 1 ,
which yields the assertion (26). By summing (26) over k from k = 0 to k = l for any l < k ˜ n δ , we have
c 1 k = 0 l μ n , k δ ( s n , k δ Y p + σ ε n , k ) r p D ξ n δ Θ ( x ^ , x n δ ) D ξ n , l + 1 δ Θ ( x ^ , x n , l + 1 δ ) + 3 k = 0 l ε n , k .
In view of Assumption 2 ( c ) and the definition of μ n , k δ , we have
μ n , k δ c 2 ( s n , k δ Y p + σ ε n , k ) 1 r p with c 2 : = min { μ 0 B ^ p , μ 1 } ,
which together with (22) and (23) gives
μ n , k δ ( s n , k δ Y p + σ ε n , k ) r p c 2 ( s n , k δ Y p + σ ε n , k ) c 2 γ p y δ F ( x n δ ) Y p c 2 γ p τ p δ p .
By inserting (31) into (30), there holds
c 1 c 2 γ p τ p δ p ( l + 1 ) D ξ n δ Θ ( x ^ , x n δ ) + 3 k = 0 l ε n , k < , 0 l < k ˜ n δ ,
which implies that k ˜ n δ < . By taking l = k n δ 1 ( k n δ = min { k ˜ n δ , k max } ) in (30), we can obtain
c 1 k = 0 k n δ 1 μ n , k δ ( s n , k δ Y p + σ ε n , k ) r p D ξ n δ Θ ( x ^ , x n δ ) D ξ n + 1 δ Θ ( x ^ , x n + 1 δ ) + 3 k = 0 k n δ 1 ε n , k ,
which yields assertion (iii).
In view of (16) and (11), we have x 0 x X ρ , i.e., x is a solution of (1) in B ρ ( x 0 ) . Then, by taking (26) with x ^ = x , we can inductively deduce that
D ξ n , k δ Θ ( x , x n , k δ ) D ξ 0 Θ ( x , x 0 ) + 3 n = 0 k = 0 k n δ 1 ε n , k .
Using (11) and (24), together with (16), we can obtain
c 0 x n , k δ x X p 2 D ξ n , k δ Θ ( x , x n , k δ ) + 2 ε n , k 2 D ξ 0 Θ ( x , x 0 ) + 8 n = 0 k = 0 k n δ 1 ε n , k 1 2 c 0 ρ p + 1 2 c 0 ρ p = c 0 ρ p ,
which gives x n , k δ x X ρ , and thus x n , k δ x 0 X 2 ρ , i.e., x n , k δ B 2 ρ ( x 0 ) for all 0 n < n δ and 0 k k n δ .
Finally, we prove that n δ < . By summing (32) over n from n = 0 to n = m for any m < n δ , and using (31), we can further obtain
c 1 c 2 γ p τ p δ p ( m + 1 ) c 1 c 2 n = 0 m k = 0 k n δ 1 μ n , k δ ( s n , k δ Y p + σ ε n , k ) r p D ξ 0 Θ ( x ^ , x 0 ) + 3 n = 0 k = 0 k n δ 1 ε n , k < ,
which yields n δ < .  □

3.1. Convergence Analysis

To carry out the convergence analysis of Algorithm 1, it is necessary to consider its counterpart with exact data. The algorithm for the noise-free case is reformulated as follows. Noting that the inner iteration number k n might not be unique (see Lemma 3 below), using different integer k n to update the outer iteration may lead to different iterative sequences ξ n , x n . Next, let Γ γ , μ 0 , μ 1 ξ 0 , x 0 denote the set of all possible sequences ξ n , x n generated by Algorithm 2 from ξ 0 , x 0 with k n chosen as in (35).
Algorithm 2 Generalized inexact Newton-Landweber iteration for exact data
Input: Parameters η < γ < 1 , μ 0 , μ 1 > 0 , k max 1 and σ > 0 ; a sequence of positive numbers { ε n , k } k 0 , n 0 satisfying n = 0 k = 0 ε n , k < ; the operator A F .
Initial guess: x 0 X and ξ 0 Θ x 0 ;
  Set n = 0 .
  Repeat:
    (i)
Assuming that ( ξ n , x n ) is constructed, we define the inner iterates { ( ξ n , k , x n , k ) } by setting ξ n , 0 = ξ n , x n , 0 = x n and
ξ n , k + 1 = ξ n , k + μ n , k A F * J r Y ( s n , k ) , x n , k + 1 = S ε n , k + 1 ( ξ n , k + 1 ) ,
where s n , k = y F ( x n ) A F ( x n , k x n ) and
μ n , k = μ ˜ n , k ( s n , k Y p + σ ε n , k ) 1 r p , i f F ( x n ) y , 0 , o t h e r w i s e
with
μ ˜ n , k = min μ 0 s n , k Y p ( r 1 ) A F * J r Y ( s n , k ) X * p , μ 1 .
    (ii)
Determine an integer 1 k n k max satisfying
s n , k Y p + σ ε n , k ( γ y F ( x n ) Y ) p , 0 k < k n
with given k max 1 and define
ξ n + 1 = ξ n , k n and x n + 1 = x n , k n .
Set n : = n + 1 .
Until stopping criterion is satisfied.
The following lemma shows the well-definedness of ξ n , x n Γ γ , μ 0 , μ 1 ξ 0 , x 0 .
Lemma 3.
Let all the conditions in Lemma 2 hold. Then, for any sequence ξ n , x n Γ γ , μ 0 , μ 1 ξ 0 , x 0 , x n , k B 2 ρ x 0 for all n 0 and k n is well-defined. Moreover, for any solution x ^ of (1) in B 2 ρ x 0 D Θ , there hold
c 3 k = 0 k n 1 μ n , k ( s n , k Y p + σ ε n , k ) r p D ξ n Θ x ^ , x n D ξ n + 1 Θ x ^ , x n + 1 + 3 k = 0 k n 1 ε n , k
and
n = 0 k = 0 k n 1 ( s n , k Y p + σ ε n , k ) <
with c 3 : = 1 β η γ p 1 p μ 0 2 c 0 1 p 1 > 0 for all n 0 .
Proof. 
Proceeding as in the proof of Lemma 3.1 in [6], we can immediately obtain (36). Summing (36) over n from n = 0 to n = , we establish that
c 3 n = 0 k = 0 k n 1 μ n , k ( s n , k Y p + σ ε n , k ) r p D ξ 0 Θ ( x ^ , x 0 ) + 3 n = 0 k = 0 k n 1 ε n , k < .
By the definition of μ n , k , we have μ n , k ( s n , k Y p + σ ε n , k ) r p c 2 ( s n , k Y p + σ ε n , k ) , from which we can obtain (37).
Next, we show that there exists 1 k n k max such that (35) holds. If F x n = y for some n, then (34) yields μ n , k = 0 for k 0 . Then it follows from the definition of ξ n , k + 1 in (33) that ξ n , k + 1 = ξ n , k = ξ n for all k 0 . Since S ε is continuous, we have
x n , k + 1 = S ε n , k + 1 ξ n , k + 1 = S ε n , k ξ n , k = x n , k = x n , k 0 .
Therefore, (35) holds for any 1 k n k max . Now, suppose that F x n y , in view of the definition of μ n , k and (35), there holds, for 0 k < k n ,
μ n , k ( s n , k Y p + σ ε n , k ) r p c 2 γ p y F x n Y p .
Combining with (36), we can obtain
c 3 c 2 γ p k n y F x n Y p c 3 k = 0 k n 1 μ n , k ( s n , k Y p + σ ε n , k ) r p D ξ n Θ x ^ , x n D ξ n + 1 Θ x ^ , x n + 1 + 3 k = 0 k n 1 ε n , k D ξ n Θ x ^ , x n + 3 k = 0 k n 1 ε n , k < ,
which suggests that k n has an upper bound, denoted by k ˜ n . Therefore, (35) holds for any k n satisfying 1 k n min k ˜ n , k max .  □
We next show the convergence of the sequence { ( ξ n , x n ) } in Γ γ , μ 0 , μ 1 ( ξ 0 , x 0 ) . The next proposition will be useful for the forthcoming convergence analysis.
Proposition 1.
Let all the conditions in Lemma 2 hold. Then, for any solution x ^ of (1) in B 2 ρ x 0 D θ , for any { ( ξ n , x n ) } Γ γ , μ 0 , μ 1 ( ξ 0 , x 0 ) , the sequence { D ξ n Θ x ^ , x n } is convergent.
Proof. 
Please refer to [26] (Lemma 3.3) for detailed proof.  □
Theorem 1.
Let X be reflexive and Y be uniformly smooth. Assume that Assumption 1, 2 and 3 hold. Then, for any sequence { ( ξ n , x n ) } Γ γ , μ 0 , μ 1 ( ξ 0 , x 0 ) , there is a solution x * of (1.1) in B 2 ρ ( x 0 ) D ( Θ ) such that
lim n x n x * X = 0 a n d lim n D ξ n Θ ( x * , x n ) = 0 .
Moreover, there holds x * = x .
Proof. 
We first prove that { x n } has a convergent subsequence. By using (38), (40) and the fact k n 1 , we have
c 3 c 2 γ p n = 0 y F x n Y p D ξ 0 Θ x ^ , x 0 + 3 n = 0 k = 0 k n 1 ε n , k < .
Consequently, lim n F ( x n ) y Y = 0 . If F ( x n ) = y for some n, by using the argument in deriving (39) repeatedly, we can show that F ( x m ) = y for all m n . Therefore, we can find a strictly increasing subsequence { n l } such that n 0 = 0 , and n l with l 0 being the first integer satisfying
n l n l 1 + 1 and F ( x n l ) y Y F ( x n l 1 ) y Y .
For such a sequence { n l } , it is easy to see that
F ( x n l ) y Y F ( x n ) y Y , 0 n n l .
Next, we show that { x n l } is a Cauchy sequence. For any solution x ^ of (1) in B 2 ρ x 0 D θ , by the definition of ε -Bregman distance, we have, for 0 j < l <
D ξ n j Θ ( x n l , x n j ) = D ξ n j Θ ( x ^ , x n j ) D ξ n l Θ ( x ^ , x n l ) + ξ n l ξ n j , x n l x ^ X * , X + ε n l .
From Proposition 1, one can see that D ξ n j Θ ( x ^ , x n j ) D ξ n l Θ ( x ^ , x n l ) tends to zero as l , j . We next estimate the term ξ n l ξ n j , x n l x ^ X * , X for 0 j < l < . By using the fact that ξ n = ξ n , 0 and ξ n + 1 = ξ n , k n , we have
ξ n l ξ n j , x n l x ^ X * , X = n = n j n l 1 ξ n + 1 ξ n , x n l x ^ X * , X = n = n j n l 1 ξ n , k n ξ n , 0 , x n l x ^ X * , X .
By the definition of ξ n , k , μ n , k and the property of J r Y , we further have
ξ n , k n ξ n , 0 , x n l x ^ X * , X = k = 0 k n 1 ξ n , k + 1 ξ n , k , x n l x ^ X * , X = k = 0 k n 1 μ n , k J r Y ( s n , k ) , A F ( x n l x ^ ) Y * , Y k = 0 k n 1 μ n , k s n , k Y r 1 A F ( x n l x ^ ) Y k = 0 k n 1 μ 1 ( s n , k Y p + σ ε n , k ) 1 1 p A F ( x n l x ^ ) Y .
Using (41), (35) and Assumption 2(c), we have
A F ( x n l x ^ ) Y A F ( x n l x n ) Y + A F ( x n x ^ ) Y ( 1 + η ) ( F ( x n l ) F ( x n ) Y + F ( x n ) y Y ) ( 1 + η ) ( F ( x n l ) y Y + 2 F ( x n ) y Y ) 3 ( 1 + η ) F ( x n ) y Y 3 ( 1 + η ) γ ( s n , k Y p + σ ε n , k ) 1 p
for 0 n n l and 0 k < k n . By inserting the above inequality into (44), there holds
ξ n , k n ξ n , 0 , x n l x ^ X * , X 3 ( 1 + η ) μ 1 γ k = 0 k n 1 ( s n , k Y p + σ ε n , k ) .
Combining with (43), we can deduce that
| ξ n l ξ n j , x n l x ^ X * , X | 3 ( 1 + η ) μ 1 γ n = n j n l 1 k = 0 k n 1 ( s n , k Y p + σ ε n , k ) .
Furthermore, together with (37), it follows that
lim j sup l j | ξ n l ξ n j , x n l x ^ X * , X | = 0 .
Since lim n ε n = 0 , we can conclude from (42) that
D ξ n j Θ ( x n l , x n j ) 0 as j , l .
By the p-convexity of Θ , there also holds x n l x n j X 0 as j , l . Therefore, { x n l } is a Cauchy sequence in X . There exists some x * X such that x n l x * as l . Please note that lim n F ( x n ) y Y = 0 , it follows from the continuity of F that F ( x * ) = y .
We next prove that x * B 2 ρ x 0 D Θ . Due to { x n l } B 2 ρ x 0 , we must have x * B 2 ρ x 0 . By ξ n l ε n l Θ ( x n l ) , there holds
Θ ( x n l ) Θ ( x ^ ) + ξ n l , x n l x ^ X * , X + ε n l .
Using (45) and x n l x * , we can find a constant C 0 such that
| ξ n l ξ n 0 , x n l x ^ X * , X | C 0 and | ξ n 0 , x n l x ^ X * , X | C 0 , l .
Therefore, | ξ n l , x n l x ^ X * , X | 2 C 0 for all l. Since Θ is lower semicontinuous and lim n ε n = 0, we can derive from (46) that
Θ ( x * ) lim inf l Θ ( x n l ) Θ ( x ^ ) + 2 C 0 < ,
which implies that x * D Θ . Thus, x * B 2 ρ x 0 D Θ is a solution of (1).
Finally, we prove the convergence of the whole sequence { x n } to x * . Let
η 0 : = lim n D ξ n Θ ( x * , x n ) ,
whose existence is guaranteed by Lemma 1. By the non-negativity of ε -Bregman distance, we have η 0 0 . Using (42) with x ^ = x * and taking l , we have
D ξ n j Θ ( x * , x n j ) D ξ n j Θ ( x * , x n j ) η 0 + sup l j | ξ n l ξ n j , x n l x * | X * , X ,
which suggests that
η 0 sup l j | ξ n l ξ n j , x n l x * | X * , X
for all j. By virtue of (45) and taking j , we obtain η 0 0 . Therefore, η 0 = 0 , i.e., lim n D ξ n Θ ( x * , x n ) = 0 . Using the p-convexity of Θ and lim n ε n = 0 , we can further conclude that lim n x n x * X = 0 .
It remains to be shown x ^ = x . From the definition of ξ n , we observe that
ξ n + 1 ξ n R ( A F * ) N ( A F ) = R ( A F ) ¯ .
Therefore, we can make use of the second part of [28] (Proposition 3.6) to complete the proof.  □

3.2. Regularization Property

In this subsection, we will prove the regularization property of Algorithm 1. Before proceeding further, the stability results of Algorithm 1 are presented in the following two lemmas: Lemma 4 is the stability property of the inner scheme, and Lemma 5 concerns the stability of the whole algorithm.
Lemma 4.
Let all the conditions in Lemma 2 hold. The sequence of noisy data { y δ l } satisfies y δ l y Y δ l 0 as l . The sequence { ( ξ n δ l , x n δ l ) } is produced by Algorithm 1. For any integer n 0 , if
ξ n δ l ξ n a n d x n δ l x n a s l
for some ( ξ n , x n ) X * × X , then there holds for each k = 0 , 1 , ,
ξ n , k δ l ξ n , k a n d x n , k δ l x n , k a s l .
with { ( ξ n , k , x n , k ) } defined by (33).
Proof. 
We use an induction argument on k. When k = 0 , the result automatically holds since ξ n , 0 δ l = ξ n , 0 = ξ n and x n , 0 δ l = x n , 0 = x n . Assume that the result holds for some k 1 , we will show that it also holds for k + 1 . The following two cases will be considered:
(i) s n , k = 0 . By the definition of ξ n , k + 1 , we have ξ n , k + 1 = ξ n , k . Consequently,
ξ n , k + 1 δ l ξ n , k + 1 = ξ n , k δ l ξ n , k + μ n , k δ l A F * J r Y ( s n , k δ l ) .
Together with μ n , k δ μ 1 ( s n , k δ Y p + σ ε n , k ) 1 r p and the property of J r Y , we can deduce that
ξ n , k + 1 δ l ξ n , k + 1 X * ξ n , k δ l ξ n , k X * + μ 1 B ^ ( s n , k δ l Y p + σ ε n , k ) 1 r p s n , k δ l Y r 1 .
By the induction hypothesis and the continuity of F, there holds ξ n , k + 1 δ l ξ n , k + 1 as l . The definition of x n , k + 1 δ l ( x n , k + 1 δ l = S ε n , k + 1 ( ξ n , k + 1 δ l ) ) and the continuity of S ε yield that x n , k + 1 δ l x n , k + 1 as l .
(ii) s n , k 0 . We first show that μ n , k δ l μ n , k as l . Recall that
μ n , k = min μ 0 s n , k Y p ( r 1 ) A F * J r Y ( s n , k ) X * p , μ 1 ( s n , k Y p + σ ε n , k ) 1 r p
and
μ n , k δ l = min μ 0 s n , k δ l Y p ( r 1 ) A F * J r Y ( s n , k δ l ) X * p , μ 1 ( s n , k δ l Y p + σ ε n , k ) 1 r p .
If A F * J r Y ( s n , k ) = 0 , we have μ n , k = μ 1 ( s n , k Y p + σ ε n , k ) 1 r p and μ n , k δ l = μ 1 ( s n , k δ l Y p + σ ε n , k ) 1 r p for sufficiently large l, which gives μ n , k δ l μ n , k as l . If A F * J r Y ( s n , k ) 0 , it follows from the induction hypothesis that μ n , k δ l μ n , k as l . Then, using the induction hypotheses and the continuity of F, J r Y and S ε , we can conclude that ξ n , k + 1 δ l ξ n , k + 1 and x n , k + 1 δ l x n , k + 1 as l .  □
Lemma 5.
Let all the conditions in Lemma 2 hold. The sequence of noisy data { y δ l } satisfies y δ l y Y δ l 0 as l . The sequence { ( ξ n δ l , x n δ l ) } is generated by Algorithm 1. Then, for any integer n 0 , by taking a subsequence of { y δ l } if necessary, there is a sequence { ( ξ n , x n ) } Γ γ , μ 0 , μ 1 ( ξ 0 , x 0 ) such that
ξ m δ l ξ m a n d x m δ l x m a s l
for all 0 m n .
Proof. 
We use an induction to complete the proof. For n = 0 , the result is trivial again. Assume, for some n > 0 , the result is true for some sequence { ( ξ n , x n ) } Γ γ , μ 0 , μ 1 ( ξ 0 , x 0 ) . We next show that it is also valid for n + 1 . To this end, we can obtain a sequence from Γ γ , μ 0 , μ 1 ( ξ 0 , x 0 ) by redefining ξ n + 1 and x n + 1 in the sequence { ( ξ n , x n ) } and applying Algorithm 2 to generate the remaining terms. We may follow Lemma 4 to derive that
ξ n , k δ l ξ n , k , x n , k δ l x n , k as l
for k = 0 , 1 , , where { ( ξ n , k , x n , k ) } are defined by (33).
Let k n δ l be the integer used to define ξ n + 1 δ l and x n + 1 δ l . By the definition of k n δ l ( k n δ l = min { k ˜ n δ l , k max } ), we know that 1 k n δ l k max . By taking a subsequence of { y δ l } if necessary, we may assume that k n δ l takes the same integer value k n . Then, 1 k n k max and
s n , k δ l Y p + σ ε n , k ( γ y δ l F ( x n δ l ) Y ) p , 0 k < k n .
Then, by taking l and using Lemma 4, we have
s n , k Y p + σ ε n , k ( γ y F ( x n ) Y ) p , 0 k < k n .
With this choice of k n , we can redefine ξ n + 1 and x n + 1 in the sequence { ( ξ n , x n ) } by ξ n + 1 : = ξ n , k n and x n + 1 : = x n , k n . The application of Lemma 4 yields ξ n + 1 δ l ξ n + 1 and x n + 1 δ l x n + 1 as l . The proof is thus complete.  □
We are now in a position to show the regularization property of Algorithm 1.
Theorem 2.
Let X be reflexive and let Y be uniformly smooth. Let Assumption 1, 2 and 3 hold. Let β > 1 , η < γ < 1 , μ 0 > 0 , and τ > 1 be chosen such that (25) holds. Assume further that y δ is a family of noisy data satisfying y δ y Y δ 0 and let n δ be determined by (23) for each y δ . Then,
x n δ δ x X 0 and D ξ n δ δ Θ x , x n δ δ 0
as δ 0 .
Proof. 
Due to the p-convexity of Θ , it is sufficient to show lim δ 0 D ξ n δ δ Θ ( x , x n δ δ ) = 0 . Since ε n , k > 0 for all n , k , from (22) and (23), we must have n δ as δ 0 . Then, for any arbitrary but fixed integer n ^ > 0 , we have n δ > n ^ for small δ . From Lemma 2, we have
D ξ n δ δ Θ ( x , x n δ δ ) D ξ n ^ δ Θ ( x , x n ^ δ ) + 3 n = n ^ n δ 1 k = 0 k n δ 1 ε n , k .
Together with the lower semi-continuity of Θ , there holds
lim sup δ 0 D ξ n δ δ Θ ( x , x n δ δ ) Θ ( x ) lim inf δ 0 Θ ( x n ^ δ ) lim δ 0 ξ n ^ δ , x x n ^ δ X * , X + 3 n = n ^ k = 0 k n δ 1 ε n , k = Θ ( x ) Θ ( x n ^ ) ξ n ^ , x x n ^ X * , X + 3 n = n ^ k = 0 k n δ 1 ε n , k = D ξ n ^ Θ ( x , x n ^ ) + 3 n = n ^ k = 0 k n δ 1 ε n , k .
Since n ^ > 0 is arbitrary, by taking n ^ , we may use (24) and Theorem 1 to deduce that lim δ 0 D ξ n δ δ Θ ( x , x n δ δ ) = 0 . By the p-convexity of Θ , we can conclude that x n δ δ x X 0 as δ 0 .  □

4. Numerical Experiments

In this section, we provide two numerical experiments. The aim of the first one is to test the effectiveness of our Algorithm 1 in identifying non-smooth solutions of parametric identification problems. The second one is to validate the efficiency of Algorithm 1 for solving non-smooth source-term problems.

4.1. Elliptic Parameter Identification

We first consider the reconstruction of parameter c in the boundary value problem [3]
u + c u = f in Ω , u = g on Ω .
from L 2 ( Ω ) —measurements of the state u, where f H 1 ( Ω ) and g H 1 2 ( Ω ) . Let
D : = { c L 2 ( Ω ) : c c ^ L 2 ( Ω ) γ 0 for all c ^ 0 a . e . }
for some γ 0 > 0 . If c D is given, then (48) has a unique solution u : = u ( c ) . Therefore, (48) reduces to solving F ( c ) = u if we define the nonlinear operator F : L 2 ( Ω ) L 2 ( Ω ) by F ( c ) : = u ( c ) . From [3], it is known that F is Fréchet differentiable; the Fréchet derivative of F and its adjoint are given by
F ( c ) h = A ( c ) 1 ( h F ( c ) ) and F ( c ) * ω = u ( c ) A ( c ) 1 ω
for h , ω L 2 ( Ω ) and A ( c ) : H 2 ( Ω ) H 0 1 ( Ω ) L 2 ( Ω ) is defined by A ( c ) u = u + c u . According to [21], if we choose A F = F c f with given c f D , condition (c) in Assumption 2 holds.
In this experiment, we pick Ω = [ 0 , 1 ] , u 0 = 1 and u 1 = 6 . The sought parameter c t is taken to be
c t = 0 , 0 t 0.1563 , 1.5 , 0.1563 < t 0.3125 , 2.5 , 0.3125 < t 0.5469 , 1.3 , 0.5469 < t 0.7813 , 0.5 , 0.7813 < t 1 .
In addition, we take u c = 1 + 5 t and the source term f t = 1 + 5 t c t . Our aim is to reconstruct c from noisy data u δ with noise level δ = u δ u c L 2 = 0.0001 . When implementing Algorithm 1, we use the initial guess c 0 = ξ 0 = 0 and
Θ x = 1 2 β Ω x w 2 d w + T V x
with β = 1 to identify the non-smooth feature of c . We fix η = 0.01 , σ = 0.001 , k max = 500 , γ = 0.98 in (22), and τ = 1.5 in (23). To meet the condition (25), we require μ 0 < 2 1 η γ 1 + η γ τ 1 η γ 1 + η γ τ β β . Thus, we take μ 0 = 1 1 1 τ τ 1.8 1 1 1 τ τ β β and μ 1 = 10 , 000 . By dividing the interval [ 0 , 1 ] into 128 subintervals of equal length, the involved differential equations are solved by the finite difference method. The minimization problems concerning Θ are solved by the primal dual hybrid gradient (PDHG) method [29] which is terminated as long as the relative duality gap is n + 1 1.5 k + 1 1.5 .
Figure 1a,b show the reconstruction results of Algorithm 1 with A F = F c f at two different choices of c f : c f = 0 and c f = 1 . As comparisons, we also consider the inexact Newton-Landweber iteration in [6], i.e., Algorithm 1 with A F being the Fréchet derivative of F at each iteration, i.e., A F = F c n δ ; the corresponding reconstructed solution is plotted in Figure 1c. As can be seen, the reconstructions of our Algorithm 1 with different values of c f have quality comparable to the ones reconstructed by [6], while Algorithm 1 does not require the information of the Fréchet derivative at each iteration. In Figure 1d, we present the evolution of the relative errors c n δ c L 2 c n δ c L 2 c L 2 c L 2 with respect to inner iteration number n. Since our Algorithm 1 avoids the calculation of the Fréchet derivative at each iteration, the computational work is considerably reduced; inevitably, Algorithm 1 may require more iterations to execute, which, however, can be offset by the numerous advantages.

4.2. A Non-Smooth Ill-Posed Problem

Let Ω R 2 be a bounded domain with a Lipschitz boundary Ω . We consider the inverse problem of the estimation of the source term f in the non-smooth semi-linear elliptic equation
Δ u + u + = f in Ω , u = 0 on Ω
from an L 2 ( Ω ) measurement u ˜ of the state u , where u + = max { u ( x ) , 0 } for all x Ω . It is easy to see that, for each f L 2 ( Ω ) , (50) has a unique solution u : = u ( f ) H 0 1 ( Ω ) C ( Ω ¯ ) L 2 ( Ω ) ; see [30] (Theorem 4.7). If we know the sought solution f and define F ( f ) = u ( f ) , then the problem (50) reduces to the inverse problem of (1). It has been shown that F is weakly closed and is not Gâteaux differentiable at f if the measurement of the set { u ( f ) = 0 } is positive; see [17] and [30] (Proposition 3.4). In this case, the Bouligand subderivative of F at f exists, which is defined by a limit of the Fréchet derivatives of F in differentiable points. Christof introduced in [31] (Proposition) a specific Bouligand subderivative of F which states that, for f L 2 ( Ω ) , the bounded linear operator G ( f ) : L 2 ( Ω ) L 2 ( Ω ) maps h H 0 1 ( Ω ) to the unique solution v : = G ( f ) h H 0 1 ( Ω ) L 2 ( Ω ) of
Δ v + χ { u ( f ) > 0 } v = h in Ω , v = 0 on Ω .
It has been shown in [21] that if we take A F = G ( M f ) with given M f L 2 ( Ω ) , the condition (17) holds for sufficiently small ρ > 0 .
In the computation, we pick Ω : = ( 0 , 1 ) × ( 0 , 1 ) R 2 and assume that the sought solution is
f ( x 1 , x 2 ) : = max ( u ( x 1 , x 2 ) , 0 ) + [ 4 π 2 u ( x 1 , x 2 ) 2 ( 2 x 1 1 ) 2 + 2 ( x 1 1 + β ) ( x 1 β ) sin ( 2 π x 2 ) ] χ [ β , 1 β ] ( x 1 )
with β = 0.1 , where
u ( x 1 , x 2 ) : = [ ( x 1 β ) 2 ( x 1 1 + β ) 2 sin ( 2 π x 2 ) ] χ [ β , 1 β ] ( x 1 )
is the corresponding exact state. Obviously, u H 0 1 ( Ω ) H 2 ( Ω ) , together the right-hand side f satisfies (50). Since u vanishes on a set of measures 2 β , the forward operator F is not Gâteaux differential at f . To carry out the computation, we employ a uniform triangular Friedrichs–Keller triangulation with 128 × 128 . The differential Equations (50) and (51) will be discretized using a finite-element method; see [17]. The corresponding discrete system is then solved by a semi-smooth Newton iteration [32]. We generate noisy data u δ by adding Gaussian noise to u with the noise level δ = u δ u L 2 . In the following, we pick δ = 0.001 and the initial guess
f 0 = f 20 × sin ( π x 1 ) sin ( 2 π x 2 ) .
When executing Algorithm 1, we use Θ x = 1 2 Ω x w 2 d w and γ = 0.9 , μ 0 = 0.8 , μ 1 = 1000 , k max = 300 , τ = 1.4 .
In Figure 2, we report the reconstruction of Algorithm 1 corresponding to A F = G ( M f ) with two different choices of M f : M f = f 0 and M f = 0 . Figure 2a displays the exact solution f . The reconstruction results of Algorithm 1 are presented in Figure 2c,d. Observe that our algorithm can provide satisfactory results for non-smooth inverse problems.
In summary, we can see from the above experiments that our Algorithm 1 could efficiently deal with smooth as well as non-smooth inverse problems by choosing the operator A F appropriately; when the forward mapping is not Gâteaux differentiable, the operator A F , as a replacement of the nonexisting Gâteaux derivative, can be taken as the Bouligand subderivative, which extends the feasibility of inexact Newton methods [4,5,6,7] for non-smooth inverse problems.

5. Conclusions

In this paper, by employing a bounded linear operator A F and inexact inner solvers, we propose a generalized inexact Newton-Landweber iteration method; this method does not require the Fréchet derivative of the forward mapping, which makes the method feasible for not only smooth but also non-smooth nonlinear inverse problems in Banach spaces. Under certain conditions on A F , the convergence analysis is carefully established. The numerical simulations on smooth parameter identification problems and non-smooth inverse source-term problems indicate that our method could effectively solve inverse problems with smooth as well as non-smooth forward operators by choosing the appropriate operator A F .
There are several possible lines of future research. First, the bounded linear operator A F may be replaced by a family of bounded linear operators A F ( x ) and the corresponding convergence theory can be developed without relying on the continuity of x A F ( x ) [17]. Second, application of the method to other non-smooth inverse problems is another interesting research direction.

Author Contributions

Conceptualization, R.G., H.F. and Z.W.; methodology, R.G., H.F. and Z.W.; formal analysis, R.G., H.F. and Z.W.; Figure creation and editing, R.G., H.F. and Z.W.; writing—original draft preparation, R.G., H.F. and Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China under grant number 42274166 and the Fundamental Research Funds for the Central Universities.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kaltenbacher, B.; Neubauer, A.; Scherzer, O. Iterative Regularization Methods for Nonlinear Ill-Posed Problems; Walter de Gruyter GmbH & Co. KG: Berlin, Germany, 2008. [Google Scholar]
  2. Scherzer, O. (Ed.) Handbook of Mathematical Methods in Imaging; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  3. Engl, H.W.; Hanke, M.; Neubauer, A. Regularization of Inverse Problems; Kluwer Academic Publishers: Boston, MA, USA; London, UK, 1996. [Google Scholar]
  4. Jin, Q. Inexact Newton-Landweber iteration for solving nonlinear inverse problems in Banach spaces. Inverse Probl. 2012, 28, 065002. [Google Scholar] [CrossRef]
  5. Margotti, F.; Rieder, A.; Leitão, A. A Kaczmarz version of the REGINN-Landweber iteration for ill-posed problems in Banach spaces. SIAM J. Numer. Anal. 2014, 52, 1439–1465. [Google Scholar] [CrossRef]
  6. Jin, Q. Inexact Newton-Landweber iteration in Banach spaces with nonsmooth convex penalty terms. SIAM J. Numer. Anal. 2015, 53, 2389–2413. [Google Scholar] [CrossRef] [Green Version]
  7. Gu, R.; Han, B. Inexact Newton regularization in Banach spaces based on two-point gradient method with uniformly convex penalty terms. Appl. Numer. Math. 2021, 160, 122–145. [Google Scholar] [CrossRef]
  8. Gu, R.; Fu, H.; Han, B. Generalized inexact Newton regularization for nonlinear ill-posed problems in Banach spaces. Inverse Probl. 2022, 38, 065001. [Google Scholar] [CrossRef]
  9. Boţ, R.I.; Hein, T. Iterative regularization with a general penalty term-theory and application to L1 and TV regularization. Inverse Probl. 2011, 28, 104010. [Google Scholar] [CrossRef] [Green Version]
  10. Zhong, M.; Wang, W.; Jin, Q. Regularization of inverse problems by two-point gradient methods in Banach spaces. Numer. Math. 2019, 143, 713–747. [Google Scholar] [CrossRef]
  11. Rieder, A. On the regularization of nonlinear ill-posed problems via inexact Newton iterations. Inverse Probl. 1999, 15, 309. [Google Scholar] [CrossRef] [Green Version]
  12. Neubauer, A. Optimal convergence rates for inexact Newton regularization with CG as inner iteration. J. Inverse Ill-Posed Probl. 2020, 28, 145–153. [Google Scholar] [CrossRef]
  13. Lechleiter, A.; Rieder, A. Towards a general convergence theory for inexact newton regularizations. Numer. Math. 2010, 114, 521–548. [Google Scholar] [CrossRef] [Green Version]
  14. Hanke, M. Regularizing properties of a truncated Newton-CG algorithm for nonlinear inverse problems. Numer. Funct. Anal. Optim. 1997, 18, 971–993. [Google Scholar] [CrossRef] [Green Version]
  15. Rieder, A. On convergence rates of inexact Newton regularizations. Numer. Math. 2001, 88, 347–365. [Google Scholar]
  16. Jin, Q. On the order optimality of the regularization via inexact Newton iterations. Numer. Math. 2012, 121, 237–260. [Google Scholar]
  17. Clason, C.; Nhu, V.H. Bouligand-Landweber iteration for a non-smooth ill-posed problem. Numer. Math. 2019, 142, 789–832. [Google Scholar] [CrossRef] [Green Version]
  18. Scherzer, O. Convergence criteria of iterative methods based on Landweber iteration for solving nonlinear problems. J. Math. Anal. Appl. 1995, 194, 911–933. [Google Scholar] [CrossRef]
  19. Clason, C.; Nhu, V.H. Bouligand-Levenberg-Marquardt iteration for a non-smooth ill-posed problem. Electron. Trans. Numer. Anal. 2019, 51, 274–314. [Google Scholar] [CrossRef] [Green Version]
  20. Mahale, P.; Dixit, S.K. Convergence analysis of simplified iteratively regularized Gauss-Newton method in a Banach space setting. Appl. Anal. 2018, 97, 2686–2719. [Google Scholar] [CrossRef]
  21. Fu, Z.; Chen, Y.; Li, L.; Han, B. Analysis of a generalized regularized Gauss-Newton method under heuristic rule in Banach spaces. Inverse Probl. 2021, 37, 125003. [Google Scholar] [CrossRef]
  22. Jin, Q. On a class of frozen regularized Gauss-Newton methods for nonlinear inverse problems. Math. Comput. 2010, 79, 2191–2211. [Google Scholar] [CrossRef] [Green Version]
  23. Zalinescu, C. Convex Analysis in General Vector Spaces; World Scientific: Singapore, 2002. [Google Scholar]
  24. Bregman, L.M. The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming. USSR Comput. Math. Math. Phys. 1967, 7, 200–217. [Google Scholar] [CrossRef]
  25. Schuster, T.; Kaltenbacher, B.; Hofmann, B.; Kazimierski, K.S. Regularization Methods in Banach Spaces; De Gruyter: Vienna, Austria, 2012. [Google Scholar]
  26. Jin, Q. Landweber-Kaczmarz method in Banach spaces with inexact inner solvers. Inverse Probl. 2016, 32, 104005. [Google Scholar] [CrossRef] [Green Version]
  27. Schöpfer, F.; Louis, A.K.; Schuster, T. Nonlinear iterative methods for linear ill-posed problems in Banach spaces. Inverse Probl. 2006, 22, 311. [Google Scholar]
  28. Jin, Q.; Wang, W. Landweber iteration of Kaczmarz type with general non-smooth convex penalty functionals. Inverse Probl. 2013, 29, 085011. [Google Scholar] [CrossRef]
  29. Zhu, M.; Chan, T.F. An efficient primal-dual hybrid gradient algorithm for total variation image restoration. UCLA CAM Rep. 2008, 34, 8–34. [Google Scholar]
  30. Tröltzsch, F. Optimal Control of Partial Differential Equations. In Theory, Methods and Applications; American Mathematical Society: Providence, RI, USA, 2010. [Google Scholar]
  31. Christof, C.; Clason, C.; Meyer, C.; Walter, S. Optimal control of a non-smooth semilinear elliptic equation. Math. Contr. Relat. Field 2018, 8, 247–276. [Google Scholar] [CrossRef] [Green Version]
  32. Ulbrich, M. Semismooth Newton Methods for Variational Inequalities and Constrained Optimization Problems in Function Spaces, (MOS-SIAM Series on Optimization); SIAM: Philadelphia, PA, USA, 2011. [Google Scholar]
Figure 1. Elliptic parameter identification ( δ = 0.0001 ). (a) Reconstruction by Algorithm 1 with A F = F c f and c f = 0 ; (b) Reconstruction by Algorithm 1 with A F = F c f and c f = 1 ; (c) Reconstruction by Algorithm 1 with A F = F c n δ ; (d) Evolution of the relative error.
Figure 1. Elliptic parameter identification ( δ = 0.0001 ). (a) Reconstruction by Algorithm 1 with A F = F c f and c f = 0 ; (b) Reconstruction by Algorithm 1 with A F = F c f and c f = 1 ; (c) Reconstruction by Algorithm 1 with A F = F c n δ ; (d) Evolution of the relative error.
Mathematics 11 01706 g001aMathematics 11 01706 g001b
Figure 2. The non-smooth ill-posed problem ( δ = 0.001 ). (a) Exact solution f ; (b) Noisy data u δ ; (c) Reconstruction by Algorithm 1 with A F = G ( M f ) and M f = 0 ; (d) Reconstruction by Algorithm 1 with A F = G ( M f ) and M f = f 0 .
Figure 2. The non-smooth ill-posed problem ( δ = 0.001 ). (a) Exact solution f ; (b) Noisy data u δ ; (c) Reconstruction by Algorithm 1 with A F = G ( M f ) and M f = 0 ; (d) Reconstruction by Algorithm 1 with A F = G ( M f ) and M f = f 0 .
Mathematics 11 01706 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gu, R.; Fu, H.; Wang, Z. Generalized Inexact Newton-Landweber Iteration for Possibly Non-Smooth Inverse Problems in Banach Spaces. Mathematics 2023, 11, 1706. https://doi.org/10.3390/math11071706

AMA Style

Gu R, Fu H, Wang Z. Generalized Inexact Newton-Landweber Iteration for Possibly Non-Smooth Inverse Problems in Banach Spaces. Mathematics. 2023; 11(7):1706. https://doi.org/10.3390/math11071706

Chicago/Turabian Style

Gu, Ruixue, Hongsun Fu, and Zhuoyue Wang. 2023. "Generalized Inexact Newton-Landweber Iteration for Possibly Non-Smooth Inverse Problems in Banach Spaces" Mathematics 11, no. 7: 1706. https://doi.org/10.3390/math11071706

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop