1,088
Views
10
CrossRef citations to date
0
Altmetric
Research Article

A high order PDE-constrained optimization for the image denoising problem

, , &
Pages 1821-1863 | Received 21 Jan 2020, Accepted 05 Dec 2020, Published online: 30 Dec 2020

Abstract

In the present work, we investigate the inverse problem of identifying simultaneously the denoised image and the weighting parameter that controls the balance between two diffusion operators for an evolutionary partial differential equation (PDE). The problem is formulated as a non-smooth PDE-constrained optimization model. This PDE is constructed by second- and fourth-order diffusive tensors that combines the benefits from the diffusion model of Perona–Malik in the homogeneous regions, the Weickert model near sharp edges and the fourth-order term in reducing staircasing. The existence and uniqueness of solutions for the proposed PDE-constrained optimization system are provided in a suitable Sobolev space. Also, an optimization problem for the determination of the weighting parameter is introduced based on the Primal–Dual algorithm. Finally, simulation results show that the obtained parameter usually coincides with the better choice related to the best restoration quality of the image.

2010 Mathematics Subject Classifications:

1. Introduction

Inverse problems are extensively studied in many areas and applications [Citation1]. Among them, the image processing and especially denoising task. The inverse problem discussed in this paper can be viewed as a variational denoising of images with total variation regularization [Citation2–4]. It is well known that the difficult choices of the weighting parameters generalize several combined first- and second-order TV-type regularization methods [Citation5,Citation6]. A major difficulty encountered in this kind of problem is how to choose the appropriate weighting parameter. In fact, there exists a vast amount of mathematical approaches on how to identify this parameter, see for example [Citation7,Citation8] and the references therein. Many of the above papers deal with the iterative and statistical approaches, which are always based on statistical estimation of the noise distribution [Citation9,Citation10]. Some successful methods are used with the total variation denoising problem, where the authors incorporated the statistical choice of a spatially variable into the standard model [Citation11,Citation12]. Other choice of the weighting parameter is based on the geometry of the image and also a specified threshold. In [Citation13], the authors propose a new algorithm which determine the weighting parameter λ based on the maximum of the Signal to Noise Ratio (SNR) of the denoised image. Similar work is also proposed in [Citation14] with satisfactory result. For a more detailed overview of methods for selecting the weighting parameter λ in general inverse problems, see [Citation15] and references therein. In a recent work, a new choice of the weighting parameter is proposed based on the learning of the noise model through a nonsmooth PDE-constrained optimization model [Citation16]. This approach is adopted for several noise distributions and has shown promising results for determining an optimal weighting parameter in the fidelity term and may be used for other variational image regularization. This weighting parameter is computed based on a training set of original and noisy images, while no information is required concerning the type or strength of the noise. In the last decade, the combined first- and high order regularizations have given an interesting success in several image processing tasks, see [Citation6,Citation17–19]. These regularizations have been introduced to reduce the staircasing effect. However, the selection of the weighting parameter λ that controls the balance between the first- and high order terms is always delicate. Indeed, an automatic selection of this parameter is desired. Recently, the bilevel techniques have had great success in this direction. The bilevel optimization is well known and understood in machine learning. This technique is considered as a semi-supervised learning method because of the ability of adapting itself to a given training set and valuable solutions. For instance, in [Citation20], the bilevel optimization was proposed for finite dimensional Markov field models. Optimal inversion and exploratory acquisition setup are studied in terms of optimal model design [Citation21]. Lately, the image processing has known the introduction of parameter learning for functional variational regularization models by several works [Citation16,Citation22–26]. One of the very interesting contributions is [Citation27] where a spatially varying regularization parameter in total variation image denoising is computed. Before that, a weighted total variation model is also investigated with bilevel optimization strategy [Citation28,Citation29]. More recently and using the same principle, a bilevel minimization framework for an automated selection of the regularization parameter in the total generalized variation (TGV) case is introduced with efficient results. However, for our knowledge, the bilevel techniques are not yet studied before for the PDE denoising models.

To avoid the problem of the blurring effect in the homogeneous regions and also to keep safe the coherence-enhancing property of the image, we propose a new PDE with two diffusion operators. This PDE takes in consideration the advantage of Perona–Malik diffusion behaviour in flat regions [Citation30], the efficiency of Weickert filter effect [Citation31–33] near sharp edges and also the ability of the fourth-order operator in avoiding the staircasing effect. Thus, taking into account these proprieties, the proposed PDE enhances the local information in images much better. However, the construction of this PDE using a convex combination between two operators requires the introduction of a weighting parameter λ=λ(x). This parameter affects the contrast of the restored image as well as the smoothness degree. In general, the choice of this parameter is not easy and it is always based on the image properties, the noise type and its nature. In this paper, we propose a PDE-constrained optimization model that compute the weighting parameter λ as well as the denoised image X without any knowledge of the image features and the noise type. In fact, we propose the study of the following optimal problem (where the attached to data term is considered to be the L1 norm since it is robust against impulse noise, see [Citation2] for more details) (1) minλUadJ(λ),(1) where J(λ)=Ω|X(T,x)Y(x)|dx+βTV(λ) and  X is a solution of the following problem: (2) Xt+2.λD(Jρσ(X))2X.(1λ)D(Jρσ(X))X=0 in ]0,T[×Ω,D(Jρσ(X))X,ν=0 on ]0,T[×Ω,X(t,x)=0, on (0,T)×Ω,X(0,x)=X0(x), in Ω,(2) where Y is the filtered image of the noisy one X0 using a bilateral filter [Citation34], ν is the exterior normal and X(T,x) is the restored image at the time T. With TV(λ)=Ω|λ|,and β is the regularization parameter. D is an anisotropic diffusion tensor and Jρ is the Structure tensor defined by (3) Jρσ(X)=Kρ(XσXσ).(3) where Xσ is constructed using a convolution of X with a Gaussian kernel. While Kρ and Kσ represent two Gaussian convolution kernels such as Kτ(x)=12πτ2exp(|x|22τ2). The function D is calculated using the tensor Jρ eigenvalues and the eigenvectors as follows: (4) D:=φ+(υ+,υ)θ+θ++φ(υ+,υ)θθ=(θ+θ)φ+(υ+,υ)00φ(υ+,υ)(θ+θ)T,(4) where υ+/ and θ+/ are respectively the eigenvalues and the eigenvectors of the tensor structure Jρ, the eigenvalues υ+/ are calculated as (5) υ+/=12trace(Jρσ)±trace2(Jρσ)4det(Jρσ).(5) The functions φ+ and φ represent the isotropic or anisotropic diffusion on image regions. Recently, an efficient choice of these functions is introduced in [Citation35], where the authors consider the behaviour of the Weickert model according to the two directions θ+ and θ. The proposed coefficients take into account the diffusion in the vicinity of the contours and corners where the eigenvalues υ+ and υ are very high. This choice is proposed as follows: (6) φ+(υ+,υ)=expυ+k1,φ(υ+,υ)=expυk21expυ+k1,(6) where k1 and k2 are two thresholds defining the diffusion with respect to the directions θ+ and θ, respectively.

We recall that λ is the weighting parameter to be estimated in (Equation1) and Uad is the set of admissible functions defined by Uad={λL1(Ω),0<λ_λλ¯<1a.e.xΩandTV(λ)C0},with λ_,λ¯ and C0 are positive constants.

Before resolving the control problem (Equation1), we have to check the existence and uniqueness of a solution of the state problem (Equation2). In order to establish the existence of the solution X to the problem (Equation2), we introduce the following set of hypotheses:

H1

– The tensor DC(R2×2,R2×2) is coercive and positive-definite matrix.

H2

– The initial condition verifies X0L(Ω).

The use of the proposed PDE-constrained with the right weighting parameter identification can cope with the famous total generalized variation denoising method [Citation36,Citation37] in the image reconstruction quality. Such a problem is considered in the framework of optimization which is challenging and generally difficult to solve, since:

  • The weighting parameter λ depends on the nature of the data, the noise level and type.

  • The nonlinear dependence of the observed noised image with respect to the noise.

  • The large computational cost due to the determination of the denoised image and the weighting parameter in the same time in the presence of numerous operators of first and high orders.

Our contribution is then to take into consideration these difficulties and provide a constraint denoising model that increases the quality of the restored image and estimates the weighting parameter λ. Moreover, inspired by the success of the Primal–Dual algorithm in accelerating the resolution of convex optimization problems [Citation38–41], we propose a new Primal–Dual algorithm adopted for PDE-constrained problems.

The rest of this paper is organized as follows: in Section 2, the variational formulation and estimations are given. Then, the existence and uniqueness results of the state problem and the optimal one are given in Sections 3 and 4 respectively. After, Section 5 presents the suggested Primal–Dual algorithm. Finally, in Section 6, we illustrated the benefits of the proposed approach by using comparative experiments with some state-of-art denoising methods.

2. Variational formulation and a priori estimates

In this section, we give the variational formulation of the problem (Equation2) and a priori estimates of the solution X. Indeed, the variational formulation of the problem (Equation2) is stated as follows: (7) FindXL2(0,T;H02(Ω))andXtL2(0,T;H2(Ω))such thatXt,ϕH2(Ω),H02(Ω)+ΩλD(Jρσ(X))2X2ϕdx+Ω(1λ)D(Jρσ(X))Xϕdx=0,ϕH02(Ω).(7) For more details about the space H02(Ω), see Appendix 2.

In order to show the existence of a solution to the problem (Equation7), we use Schauder fixed point Theorem [Citation42]. For this reason, we need to show a priori estimates of the solution to the problem (Equation7) in L2(0,T;H02(Ω)).

Lemma 2.1

Assume that assumptions (H1)(H2) are satisfied, then there exists three constants Ci>0 for i = 1, 2, 3, such that the weak solution of the problem (Equation7) satisfies the following a priori estimations: (8) XL(0,T;L2(Ω))C1,(8) (9) XL2(0,T;H02(Ω))C2,(9) (10) XtL2(0,T;H2(Ω))C3.(10)

Proof.

Consider τ(0,T], then for t(0,τ) we take ϕ=X(t) in the variational formulation (Equation7), we get: 12ddtX(t)L2(Ω)2dt+ΩλD(Jρσ(X))|2X|2dx+Ω(1λ)D(Jρσ(X))|X|2dx=0,we integrate with respect to t over (0,τ) (11) 12X(τ)L2(Ω)2+0τΩλD(Jρσ(X))|X2|2dxdt+0τΩ(1λ)D(Jρσ(X))|X|2dxdt=12X(0)L2(Ω)2.(11) Since λUad, we find 12X(τ)L2(Ω)2+λ_0τΩD(Jρσ(X))|2X|2dxdt+(1λ¯)0τΩD(Jρσ(X))|X|2dxdt12X(0)L2(Ω)2.On the other hand, we have D(Jρσ) is coercive, where the coercivity constant is α, then 12X(τ)L2(Ω)2+αλ_0τΩ|2X|2dxdt+α(1λ¯)0τΩ|X|2dxdt12X(0)L2(Ω)2,which implies that X(τ)L2(Ω)2C1,τ[0,T],where the constant C1=X(0)L2(Ω)2. Consequently, we have (12) XL(0,T;L2(Ω))2C1,XL2(0,T;L2(Ω))2TXL(0,T;L2(Ω))2.(12) We have also (13) min12T,αλ_,(1λ¯)αXL2(0,T;H2(Ω))2C1,(13) Since we have XL2(0,T;H02(Ω))2XL2(0,T;H2(Ω))2, then XL2(0,T;H02(Ω))2C2,with C2=12C1min(12T,αλ_,(1λ¯)α).Let's prove now the estimation of Xt. From the variational formulation (Equation7), we have (14) Xt,ϕH2(Ω),H02(Ω)=|ΩλD(Jρσ(X))2X2ϕdx+Ω(1λ)D(Jρσ(X))Xϕdx|,ϕL2(0,T;H02(Ω)).(14) Using the Hölder inequality and the fact that D(Jρσ(X)) is bounded in L(0,T,C(Ω)), we find (15) Xt,ϕH2(Ω),H02(Ω)λ¯C2XL2(Ω)2ϕL2(Ω)+(1λ_)CXL2(Ω)ϕL2(Ω),ϕL2(0,T;H02(Ω)).(15) or 2ϕL2(Ω)ϕH02(Ω) and ϕL2(Ω)ϕH02(Ω), we obtain (16) supϕH02(Ω)|Xt,ϕH2(Ω),H02(Ω)|ϕH02(Ω)λ¯C2XL2(Ω)+(1λ_)CXL2(Ω).(16) Using norm equivalence and Poincaré inequality we have 2XL2(Ω)XH2(Ω)CXH02(Ω) (see Appendix 2), XL2(Ω)CXH02(Ω) and by integrating on t(0,T], we obtain (17) XtL2(0,T,H2(Ω))CC2=C3.(17) This achieves the proof.

3. Existence and uniqueness of the state problem

In this section, based on the above estimates, we will show the existence and uniqueness of the solution for the problem (Equation7) for a fixed λUad. The main difficulties encountered in this study come from the strong nonlinearity of the conductivity part of Equation (Equation2). To overcome these difficulties, we use the Schauder fixed point [Citation42]. The idea of the proof is inspired from the paper [Citation43]. In order to prove the existence of a solution to the problem (Equation7) we will prove some lemmas that will be useful for the theorem of the existence. Let's now define the following operator G by (18) G :VVX¯X,(18) where X is the unique solution of (19) withXL2(0,T;H02(Ω))andXtL2(0,T;H2(Ω))such thatXt,ϕH2(Ω),H02(Ω)+ΩλD(Jρσ(X¯))2X2ϕdx+Ω(1λ)D(Jρσ(X¯))Xϕdx=0,ϕH02(Ω),(19) and (20) V={UL2(0,T;H01(Ω))L(0,T;L2(Ω))/UL(0,T;L2(Ω))C1,UL2(0,T;H02(Ω))C2andUtL2(0,T,H2(Ω))C3}.(20) The operator G is well defined, thanks to two things: first, the existence of the solution for the problem (Equation19) by the equivalent of Lax–Milgram theorem for parabolic equation which called Lions theorem. This theorem is stated in the comments part of the Chapter 10 of [Citation44] (also it is stated in pages 509−510 of the book [Citation45]) and it is mentioned in the part of parabolic equation from the book [Citation46]. Second, using the same techniques as in the proof of Lemma 2.1, it's easy to prove that for all X¯V, we have XV.

To prove that G admits a unique fixed point X, which is the solution of (Equation7). We have to show that G is continuous and compact. The following lemma show the continuity of the operator G.

Lemma 3.1

The operator G is continuous in V.

Proof.

In order to prove the continuity of the map G, let us consider the sequence X¯n in L2(0,T,H01(Ω))L(0,T;L2(Ω)), such that (21) X¯nnX¯inL2(0,T,H01(Ω))L(0,T;L2(Ω)),(21) we have to prove that (22) Xn=G(X¯n)nX=G(X¯)inL2(0,T,H01(Ω))L(0,T;L2(Ω)).(22) Let Xn (respectively X) be the unique solution of X¯n (respectively X¯) for the formulation Equation19, which means (23) Xnt,ϕH2(Ω),H02(Ω)+ΩλD(Jρσ(X¯n))2Xn2ϕdx+Ω(1λ)D(Jρσ(X¯n))Xnϕdx=0,ϕL2(0,T;H02(Ω)),(23) (24) Xt,ϕH2(Ω),H02(Ω)+ΩλD(Jρσ(X¯))2X2ϕdx+Ω(1λ)D(Jρσ(X¯))Xϕdx=0,ϕL2(0,T;H02(Ω)).(24) Using Equations (Equation23) and (Equation24), and by taking ϕ=XnX, we find (XnX)t,XnXH2(Ω),H02(Ω)+ΩλD(Jρσ(X¯))2(XnX)2(XnX)dx+Ω(1λ)D(Jρσ(X¯))(XnX)(XnX)dx=Ωλ(D(Jρσ(X¯))D(Jρσ(Xn¯)))2Xn2(XnX)dx+Ω(1λ)(D(Jρσ(X¯))D(Jρσ(Xn¯)))Xn(XnX)dx.Integrating now on t(0,τ],τ(0,T] and using the fact that the operator D(Jρσ) is smooth enough, we have (25) D(Jρσ(X¯n))D(Jρσ(X¯))L(0,τ;L(Ω))CX¯nX¯L(0,τ;L2(Ω)).(25) Using the fact that D is coercive, and using the Hölder inequality, we obtain (26) 12Xn(τ)X(τ)L2(Ω)2+λ_α2(XnX)L2(0,τ;L2(Ω))2(26) (27) +(1λ¯)α(XnX)L2(0,τ;L2(Ω))2λ¯CX¯nX¯L(0,τ;L2(Ω))2XnL2(0,T;L2(Ω))2(XnX)L2(0,τ;L2(Ω))+(1λ_)X¯nX¯L(0,τ;L2(Ω))XnL2(0,T;L2(Ω))(XnX)L2(0,τ;L2(Ω)).(27) Using the Young inequality for the right terms, we find (28) 12Xn(τ)X(τ)L2(Ω)2+λ_αϵ1Cλ¯22(XnX)L2(0,τ;L2(Ω))2+(1λ¯)αϵ1C(1λ_)2(XnX)L2(0,τ;L2(Ω))2λ¯C2ϵ1X¯nX¯L(0,τ;L2(Ω))22XnL2(0,T;L2(Ω))2+(1λ_)C2ϵ1X¯nX¯L(0,τ;L2(Ω))2XnL2(0,T;L2(Ω))2,(28) and using the Poincaré inequality: 2XnL2(0,T;L2(Ω))2CXnL2(0,T;H02(Ω))2,XnL2(0,T;L2(Ω))2CXnL2(0,T;H02(Ω))2,we obtain (29) 12Xn(τ)X(τ)L2(Ω)2+λ_αϵ1Cλ¯22(XnX)L2(0,τ;L2(Ω))2(29) (30) +(1λ¯)αϵ1C(1λ_)2(XnX)L2(0,τ;L2(Ω))2λ¯C2ϵ1+(1λ_)C2ϵ1X¯nX¯L(0,τ;L2(Ω))2XnL2(0,T;H02(Ω))2.(30) For ϵ1<min(2(1λ¯)αC(1λ_),2λ_αCλ¯) chosen and since we have (31) X¯nX¯inL(0,T;L2(Ω)),(31) we find that (32) XnXinL2(0,T,H02(Ω))L(0,T,L2(Ω)).(32) By the equivalence between the norms .H02(Ω) and .H2(Ω) (see Appendix 2), we find that the operator G is continuous.

Let's show now that G is compact.

Lemma 3.2

The operator G is compact in V.

Proof.

Now, let us show that G is compact. For this, let (X¯n)n be a bounded sequence in L2(0,T;H01(Ω))L(0,T;L2(Ω)), and let Xn=G(X¯n) be the unique solution of (Equation19) to X¯n. Indeed, by taking ϕ=Xn and integrating over t and using the same proof as in Lemma 2.1, we have (33) 12Xn(τ)L2(Ω)2+αλ_2XnL2(0,τ;L2(Ω))2+(1λ¯)αXnL2(0,τ;L2(Ω))212C1,τ0,T.(33) We have also (34) XntL2(0,τ;H2(Ω))2C3.(34) Now, thanks to the compact embedding of H02(Ω) in H01(Ω) and the continuous embedding of H01(Ω) in H2(Ω). From Aubin–Lions Lemma [Citation45–47], we have uL2(0,T;H02(Ω))/utL2(0,T;H2(Ω))is compactly embedded in L2(0,T;H01(Ω)), which means we can extract a subsequence denoted again (Xn)n that converges in L2(0,T;H01(Ω)). This implies that the operator G is compact.

The following theorem shows the existence of weak solution for the problem (Equation7)

Theorem 3.1

The problem (Equation7) admits a unique solution in L2(0,T;H01(Ω)).

Proof.

Existence We have that the operator G is well defined. From Lemmas 3.1 and 3.2, G is continuous and compact in V, which implies the existence of the fixed point by using the Schauder fixed point [Citation42].

Uniqueness To prove the uniqueness of the solution, we assume that X1 and X2 are two distinct solutions of the problem (Equation7). By subtracting the weak formulations of the solutions X1 and X2, we obtain (35) X1tX2t,ϕH2(Ω),H02(Ω)+Ωλ(D(Jρσ(X1))2X1D(Jρσ(X2))2X2)2ϕdx+Ω(1λ)(D(Jρσ(X1))X1D(Jρσ(X2))X2)ϕdx=0,(35) taking ϕ=X1X2, we get (36) X1tX2t,X1X2H2(Ω),H02(Ω)+Ωλ(D(Jρσ(X1))2X1D(Jρσ(X2))2X2)2(X1X2)dx+Ω(1λ)(D(Jρσ(X1))X1D(Jρσ(X2))X2)(X1X2)dx=0.(36) Then, we have X1tX2t,X1X2H2(Ω),H02(Ω)+ΩλD(Jρσ(X1))2(X1X2)2(X1X2)dx+Ω(1λ)D(Jρσ(X1))(X1X2)(X1X2)dx=Ωλ(D(Jρσ(X1))D(Jρσ(X2)))2X22(X2X1)+Ω(1λ)(D(Jρσ(X1))D(Jρσ(X2)))X2(X2X1).By using the fact that D is bounded and by the Hölder inequality, we find d2dtX1(t)X2(t)L2(Ω)2+αλ_2(X1X2)L2(Ω)2+(1λ¯)α(X1X2)L2(Ω)2λ¯D(Jρσ(X1(t)))D(Jρσ(X2(t)))L(Ω)2X2(t)L2(Ω)2(X1(t)X2(t))L2(Ω)+(1λ_)D(Jρσ(X1(t)))D(Jρσ(X2(t)))L(Ω)X2(t)L2(Ω)(X1(t)X2(t))L2(Ω).Since the operator D(Jρ) is smooth enough, we have (37) D(Jρσ(X1(t)))D(Jρσ(X2(t)))L(Ω)CX1(t)X2(t)L2(Ω),t0,T.(37) Then (38) d2dtX1(t)X2(t)L2(Ω)2+λ_α2(X1X2)L2(Ω)2+(1λ¯)α(X1X2)L2(Ω)2λ¯CX1(t)X2(t)L2(Ω)2X2(t)L2(Ω)2(X1(t)X2(t))L2(Ω)+(1λ_)CX1(t)X2(t)L2(Ω)X2(t)L2(Ω)(X1(t)X2(t))L2(Ω).(38) Using the Young inequality, we find (39) d2dtX1(t)X2(t)L2(Ω)2+λ_αϵ1Cλ¯22(X1X2)L2(Ω)2+(1λ¯)αϵ1C(1λ_)2(X1X2)L2(Ω)2CX1(t)X2(t)L2(Ω)2λ¯12ϵ12X2(t)L2(Ω)2+(1λ_)12ϵ1X2(t)L2(Ω)2C2ϵ1X1(t)X2(t)L2(Ω)X2(t)H02(Ω)2.(39) By taking ϵ1<min(2(1λ¯)αC(1λ_),2λ_αCλ¯) and by using the Grönwall inequality, we conclude that (40) X1X2L(0,T;L2(Ω))20.(40) Finally, we get the uniqueness of the solution.

4. Existence of solution for the optimization problem

We prove now the existence of the solution to the optimal control problem (Equation1) we need the following compactness lemmas.

Lemma 4.1

Uad is compact for the topology defined by the strong convergence in L1(Ω).

Proof.

Let's (λk)k be a sequence in Uad. Then we have (41) λ_λkλ¯a.e. Ω(41) and (42) TV(λk)C0.(42) Our aim is to show that we can extract a subsequence denoted again (λk)k such that λkλinL1(Ω)and λUad.From the estimations (Equation41) and (Equation42) we have (43) λkBV(Ω)C0+λ¯mes(Ω).(43) Using the compact embedding of BV(Ω) in L1(Ω), we can extract a subsequence denoted again (λk)k such that λkλinL1(Ω) with 0<λ_λλ¯<1a.e.xΩFrom estimation (Equation42) we find λkλinMb(Ω),Mb(Ω) is the space of bounded measures, using the Banach–Alaoglu–Bourbaki theorem in [Citation44], we find TV(λ)C0.Which conclude the compactness of Uad.

Now we will prove the continuity of the map λX(λ). Which means that if (λk)k converges to λ in Uad, then XkX(λk) the solution of (Equation7) converges to XX(λ) solution of (Equation7).

Proposition 4.1

Let (λk)k be a sequence from Uad that converges in L1(Ω) to λUad and XkX(λk) solution of (Equation7) related to λk, kN, we have

(1)

The sequence (Xk)k converges weakly in L2(0,T;H02(Ω)) to XX(λ) solution of (Equation7) related to λ.

(2)

The cost functional J verifies J(λ)lim infk+J(λk).

Proof.

  1. Let's (λk)k be a sequence of Uad and XkXk(λk) solution of (Equation7) related to λk, kN, which means (44) Xkt,ϕH2(Ω),H02(Ω)+ΩλkD(Jρσ(Xk))2Xk2ϕdx+Ω(1λk)D(Jρσ(Xk))Xkϕdx=0,(44) ϕH02(Ω). From Lemma 2.1, we have: there exists Ci>0 for i{1,2,3} independent of k, such that (45) XkL(0,T;L2(Ω))C1,(45) (46) XkL2(0,T;H02(Ω))C2,(46) (47) XktL2(0,T;H2(Ω))C3.(47) Since (λk)k is a sequence of Uad and using the results of compactness of Uad obtained in Lemma 4.1, we can extract a subsequence denoted again (λk)k and λ, such that (48) λkkλinL1(Ω).(48) Also, from the boundedness of (Xk)k in L2(0,T;H02(Ω)) and Xkt in L2(0,T;H2(Ω)), we can extract a subsequence denoted again (Xk)k (49) XkXin L2(0,T;H02(Ω))(49) and (50) XktXtin L2(0,T;H2(Ω)).(50) By using Aubin–Lions Lemma [Citation45–47], we have (51) XkXin L2(0,T;H1(Ω))(51) and (52) XkXin C(0,T;H1(Ω)).(52) Thanks to the compact embedding of H01(Ω) in L2(Ω) and the continuous embedding of L2(Ω) in H2(Ω). Since, we have (Xk)k in L(0,T;H01(Ω)) and Xkt in L2(0,T;H2(Ω)), using Aubin–Simon lemma [Citation48] we obtain (53) XkXin C([0,T];L2(Ω)).(53) Let's prove now that X is the solution of (Equation7). Which means, it suffices to prove that the following convergences hold: (54) 0TXkt,ϕH2(Ω),H02(Ω)dtk0TXt,ϕH2(Ω),H02(Ω)dt,(54) (55) 0TΩλkD(Jρσ(Xk))2Xk2ϕdxdtk0TΩλD(Jρσ(X))2X2ϕdxdt(55) and (56) 0TΩ(1λk)D(Jρσ(Xk))Xkϕdxdtk0TΩ(1λ)D(Jρσ(X))Xϕdxdt.(56) For the convergence (Equation54), it is obtained directly from convergence (Equation50). For the convergences (Equation55) and (Equation56), it suffices to show only one of them since they are similar. For this reason, we will show only the convergence (Equation55). Let's consider I=0TΩλkD(Jρσ(Xk))2Xk2ϕdxdt0TΩλD(Jρσ(X))2X2ϕdxdt,using the following decomposition I=I1+I2+I3, where I1=0TΩ(λkλ)D(Jρσ(Xk))2Xk2ϕdxdt,I2=0TΩλ(D(Jρσ(Xk))D(Jρσ(X)))2Xk2ϕdxdtand I3=0TΩλD(Jρσ(X))2(XkX)2ϕdxdt.Let's show that I1k0.

    We have |I1|0TΩ|λkλ||D(Jρσ(Xk))||2Xk|2ϕ|dxdt,using Hölder inequality and the proprieties of D(Jρσ(Xk)), we find |I1|22XkL2(0,T;L2(Ω))2D(Jρσ(Xk))L(0,T;L(Ω))20TΩ(λkλ)2|2ϕ|2dxdt.Since we have D(Jρσ(Xk))L(0,T;L(Ω))2CXkL(0,T;L2(Ω))2and using the estimations (Equation45) and (Equation46) we obtain |I1|2C0TΩ(λkλ)2|2ϕ|2dxdt.On the other hand, we know that (57) λkkλinL1(Ω),(57) then (λkλ)2|2ϕ|20, a.e in ΩϕL2(0,T;H2(Ω)),and λk,λ are bounded (λkλ)2|2ϕ|24λ¯2|2ϕ|2L1(0,T;L1(Ω)).By applying the dominated convergence theorem of Lebesgue, we have I1k0.Now we will show that I2k0. Using the Hölder inequality and the fact that λ is bounded, we obtain |I2|λ¯2XkL2(0,T;L2(Ω))D(Jρσ(Xk))D(Jρσ(X))L(0,T;L(Ω))2ϕL2(0,T;L2(Ω)).Since the operator D(Jρ) is smooth enough, we have (58) D(Jρσ(Xk))D(Jρσ(X))L(0,T;L(Ω))CXkXL(0,T;L2(Ω)).(58) Using the estimation (Equation45) then |I2|CXkXL(0,T;L2(Ω))2ϕL2(0,T;L2(Ω)).By the convergence (Equation53) we obtain I2k0.We show now that I3k0. We have λ is bounded in L(Ω) and the following inequality (59) D(Jρσ(X))L(0,T;L(Ω))CXL(0,T;L2(Ω))C.(59) In other hand, we have XkXin L2(0,T;H02(Ω)),then I3k0,which means Ik0.

  2. Let's show that the cost functional J verifies J(λ)lim infk+J(λk).Indeed, we have (60) XkXin C([0,T];L2(Ω)),(60) then (61) Ω|Xk(T,x)Y(x)|dxΩ|X(T,x)Y(x)|dx.(61) Indeed, we have (62) Ω|Xk(T,x)Y(x)|dxΩ|X(T,x)Y(x)|dxΩ|Xk(T,x)X(T,x)|dxmes(Ω)12Xk(T,)X(T,)L2(Ω)Csupt[0,T]Xk(t,)X(t,)L2(Ω)CXkXC([0,T],L2(Ω)).(62) Using the convergence (Equation60) we obtain Ω|X(T,x)Y(x)|dxlim infk+Ω|Xk(T,x)Y(x)|dx.Also, we have TV(λk)C0,then we can extract a subsequence denoted again (λk), such that λkλinMb(Ω),which means TV(λ)lim infk+TV(λk).Finally, we obtain (63) J(λ)=Ω|X(T,x)Y(x)|dx+TV(λ)lim infk+Ω|Xk(T,x)Y(x)|dx+lim infk+TV(λk)lim infk+(Ω|Xk(T,x)Y(x)|dx+TV(λk))lim infk+J(λk).(63)

The following theorem shows the existence of optimal solution to the problem (Equation1).

Theorem 4.1

The problem (Equation1) admits at least a solution in Uad.

Proof.

Let's (λk)k be a minimizing sequence of J in Uad such that limkJ(λk,Xk)=infλUadJ(λ,X(λ)).Using Lemma 4.1, which means that Uad is compact, then we can extract a subsequence denoted again (λk)k that converge to λ in Uad. In other hand, from Proposition 4.1, we have J is lower semicontinuous, then: J(λ,X)lim infkJ(λk,Xk)limkJ(λk,Xk)=infλUadJ(λ,X(λ)),we have also infλUadJ(λ,X(λ))J(λ,X).Then limkJ(λk,Xk)=infλUadJ(λ,X(λ))=J(λ,X).This concludes that the problem (Equation1) admits a solution in Uad.

5. The proposed algorithm and numerical approximation

In this section, we give the proposed algorithm based on the Primal–Dual algorithm and the numerical approximation of the state and adjoint problems.

5.1. The proposed algorithm

Recently, an extension of the A. Chambolle and T. Pock Primal–Dual algorithm (proposed in [Citation38]) to nonsmooth optimization problems, which involves complex nonlinear operators coming from PDE-constrained optimization problems [Citation49]. Inspired by this work, we propose an efficient Primal–Dual algorithm, well adapted to the evolutionary PDE-constrained problem, where two operators are involved (which to our knowledge has not already been treated). Indeed, we transform the studied PDE into an optimization problem of the following form: (64) λˆ=argminλ{ F(K(λ))+G(λ)},(64) where F and G are proper, convex and lower semicontinuous functional and K is a nonlinear operator involving the solution of the PDE in (Equation2). First, we construct the operator S that maps λ to the solution X of the PDE in (Equation23) at final time T, which is defined by S : UadL2(Ω)λX(T,x),where X is the weak solution of (65) Xt,ϕH2(Ω),H02(Ω)+ΩλD(Jρσ(X))2X2ϕdx+Ω(1λ)D(Jρσ(X))Xϕdx=0,ϕH02(Ω).(65) Let's transform the problem (Equation1) to the form (Equation64). The problem (Equation1) is given as follows: (66) λˆ=argminλUad{||S(λ)Y||L1(Ω)+βλL1(Ω)},=argminλL1(Ω){||S(λ)Y||L1(Ω)+βλL1(Ω)+iUad(λ)},=argminλL1(Ω){F(K(λ))+G(λ)},(66) where the functions F and G are given as follows: F : (L1(Ω))2R{}g=g1g2gL1(Ω)=g1L1(Ω)+g2L1(Ω),and G : UadR{}λiUad(λ)=0,λUad,+,λUad.We define also the operators K1, K2 such as K1(λ)=S(λ)Y and K2(λ)=βλ. The operator K is then defined by (67) K=K1K2,(67) where K1 is nonlinear operator and K2 is linear, with these notations, we have (68) F(K(λ))=K1(λ)L1(Ω)+K2(λ)L1(Ω).(68) Now, we can apply the iterations of the Primal–Dual Algorithm [Citation38] to our problem (Equation66), we obtain the following iterative algorithm: (69) λn+1=(I+ηG)1(λnηK(λn)ϑn),λˆn+1=λn+1+θ(λn+1λn),ϑn+1=(I+δF)1(ϑn+δK(λˆn+1)),(69) where the F:(L(Ω))2R{} denotes the Legendre-Fenchel conjugate of the functional F is given by (70) F(ϑ)=0ϑ1O+ϑ1O+0ϑ2O+ϑ2O,(70) where O={Z:ZL(Ω)1}.

In order to give the Primal–Dual algorithm, we have to expression the proximal operator function (I+δF)1 and (I+τG)1 (for more details see [Citation49]). For this, we have (71) ϑ=(I+δF)1(ϑˆ)=proj(ϑˆ)=projO(ϑˆ1)projO(ϑˆ2),(71) where projO(ϑˆi)=ϑˆimax(ϑˆiL(Ω),1),for i=1,2,and (72) λ=(I+ηG)1(λˆ)=projUad(λˆ).(72) We need also to define the operator K, which is the adjoint derivative of the operators K defined by (73) K(λ)=K1,K2=S(λ),βdiv,(73) where S(λ):L2(Ω)(Uad) denotes the adjoint of the Gâteaux derivative of the operator S, it is given by the following expression: (74) S(λ)W=0TD(Jρσ(X))2X2Pdt+0TD(Jρσ(X))XPdt,(74) where W=ϑ1 in Algorithm 1 and P is the solution of the following adjoint problem: (75) Pt+2.λD(Jρσ(X))2P.(1λ)D(Jρσ(X))P+λD(Jρσ(X))2X2P+(1λ)D(Jρσ(X))XP=0 in ]0,T[×Ω,D(Jρσ(X))P,ν=0 on ]0,T[×Ω,P(t,x)=0, on (0,T)×Ω,P(T,x)=W(x), in Ω,(75) where D(Jρσ(X))=D(Jρσ(X))(Jρσ(X)) is the derivative of D(Jρσ(X)) with respect to state solution X.

For the proof of the Gâteaux derivative of the operator S and the calculation of S(λ), see Appendix 1.

After introducing all the operators needed, the proposed algorithm is given as follows:

For more details about the convergence of Algorithm 1, see [Citation49].

5.2. Numerical approximation

This part provides all the ingredients needed to implement the Primal–Dual algorithm adopted to the problem (Equation64). We discretize first the time interval [0,T] into a series of M subdomains, namely [0,T]=m=1j=M[tm1,tm],withΔt=tmtm1=TM+1.Using a semi-discretization with respect to time for the problem (Equation2) and the problem (Equation75), we have m=0,,M1 (76) Xm+1Xmdt+2.(λD(Jρσ(Xm))2Xm)+.(1λ)D(Jρσ(Xm))Xm)=0. in Ω,D(Jρσ(Xm))Xm,ν=0 on Ω,Xm(x)=0, on Ω,X0(x)=X0(x), in Ω.(76) and m=M,,1 (77) PmPm1dt+2.λD(Jρσ(Xm))2Pm.(1λ)D(Jρσ(Xm))Pm+λD(Jρσ(Xm))2Xm2Pm+(1λ)D(Jρσ(Xm))XmPm=0 in Ω,D(Jρσ(Xm))Pm,ν=0 on Ω,Pm(x)=0, on Ω,PM(x)=W(x), in Ω.(77) The discretization of the gradient operator X=((X)1,(X)2) knowing that Xi,j,i,j=1,(M1,M2) is the discrete image, and RM1×M2 the set of discrete images. (78) (X)i,j1=Xi+1,jXi,jif i<M10if i=M1,(X)i,j2=Xi,j+1Xi,jif j<M20if j=M2,(78) we also need to define the discretization of the adjoint operator of the gradient ” div ” : (RM1×M2)2RM1×M2”, satisfying the following relation: (79) divYX=YX,XRM1×M2,Y(RM1×M2)2.(79) With: (div(Y1,Y2))i,j=(div(Y1,Y2))i,j1+(div(Y1,Y2))i,j2and (80) (div(Y1,Y2))i,j1=Yi,j1Yi1,j1if 1<i<M1Yi,j1ifi=10ifi=M1,({div}(Y1,Y2))i,j2=Yi,j2Yi,j12if 1<j<M2Yi,j2if j=1Yi,j12if j=M2.(80) We define also the second-order differential operator 2=xxxyxyyy, where xx, yy, and xy are given in discrete form such as (81) xxXi,j=Xi,M22Xi,1+Xi,2if1iM1,j=1,Xi,j12Xi,j+Xi,j+1if1iM1,1<j<M2,Xi,M212Xi,M2+Xi,1if1iM1,j=M2,(81) and (82) yyXi,j=XM1,j2X1,j+X2,jifi=1,1jM2,Xi1,j2Xi,j+Xi+1,jif1<i<M1,1jM2,XM11,j2XM1,j+X1,iifi=M1,1jM2.(82) On the other hand, we have (83) xyXi,j=Xi,jXi+1,jXi,j+1+Xi+1,j+1if1i<M1,1j<M2,Xi,M2Xi+1,M2Xi,1+Xi+1,1if1i<M1,j=M2,XM1,jX1,jXM1,j+1+X1,j+1ifi=M1,1j<M2,XM1,M2X1,M2XM1,1+X1,1ifi=M1,j=M2.(83) For the approximation of D(Jρσ(X))=D(Jρσ(X))(Jρσ(X)), we use the following approximation: (Jρσ(Xi,j))=Jρσ(Xi+1,j)Jρσ(Xi,j)Xi+1,jXi,jor (Jρσ(Xi,j))=Jρσ(Xi,j+1)Jρσ(Xi,j)Xi,j+1Xi,j,for i,j=1,(M11,M21).

6. Numerical results

In this section, we present numerous experiment results to see the contribution of the proposed model. In fact, the results are divided into two categories: the first one is devoted to the results of the case where λ is a scalar and the second deals with the case of λ is spatially varying. For a fair comparison, our method and the compared ones are implemented using Matlab 2013a on the platform: 3 GHz dual-core central processing unit (CPU) and 8 Gbytes RAM. The stopping criterion is the relative residual error becoming less than 104: (84) Err=λˆn+1λˆn2λˆn+12104.(84) In the following numerical experiments, we fix the values of λ_ and λ¯ (the lower and upper bounds of the parameter λ in Uad) for the proposed algorithm such as λ_=103 and λ¯=0.8.

6.1. The case of scalar λ

In this part, we are interested in the case where the parameter λ is constant. We focus on image denoising task with the aim to reduce the staircasing effect and intensity loss for different noise type. We compare our model to the classical Rudin, Osher and Fatemi (ROF) model [Citation3] solved by the Bregman iteration [Citation50], the nonlocal means (NLM) model [Citation51,Citation52], the combined TV and TV2 regularization (TV+TV2) [Citation19,Citation53] solved by Bregman iteration and the TGV model [Citation36,Citation37] solved by the Primal–Dual algorithm. Note that all parameters for the compared models were optimized in order to meet the higher Peak signal-to-noise ratio (PSNR) values.

We start with the first test, where the image of Pirate is considered. The image size is 510×510 pixels. We construct a corrupted version of the original image by adding a with Gaussian noise of zero mean and standard deviation σ=0.3. For this example, the chosen parameters (α0,α1) for the TV+TV2 and TGV methods optimized the PSNR value as we can see in Figure . Concerning the NLM model we chose the size of the neighbourhood 9×9 and the threshold that determines isotropic/anisotropic neighbourhood selection λ=22, while we take α=0.1 for the ROF model. These parameters are chosen using the same procedure for all the remaining tests.

Figure 1. Plot of the PSNR values of the restored image using the TGV and TV+TV2 approach with respect to the parameters α0 and α1. We can see that the highest achieved PSNR for the TGV2 approach is related to the values α0=0.06 and α0=0.04. For the TV+TV2 model, the best PSNR corresponds to the values α0=0.011 and α0=0.05. Note that the original image is the Pirate one corrupted by Gaussian noise of variance 0.3.

Figure 1. Plot of the PSNR values of the restored image using the TGV and TV+TV2 approach with respect to the parameters α0 and α1. We can see that the highest achieved PSNR for the TGV2 approach is related to the values α0=0.06 and α0=0.04. For the TV+TV2 model, the best PSNR corresponds to the values α0=0.011 and α0=0.05. Note that the original image is the Pirate one corrupted by Gaussian noise of variance 0.3.

In Figure , we show the respective denoised results by different methods. Visually, our method can achieve a reconstruction which is close to one obtained by the TGV method while the other models are outperformed. In the second denoising test, we take the Cameraman image which is a challenging textural image. We consider also the Gaussian noise but with a higher standard deviation σ=0.5. In order to see the robustness of the proposed method, we present in Figure  the restored image with different approaches including the proposed one. Once again, we can observe that visually the proposed method can achieve a better reconstructed result compared to other methods. To see the evolution of the obtained weighting parameter λ, we plot in Figure  the values of λ with respect to the iterations for the two above examples. It is then observed that the obtained parameter for the Pirate image is λ=0.39, while for the Cameraman image the parameter is λ=0.68.

Figure 2. The obtained denoised image compared to other classical approaches for the (Pirate image), where the noise is considered to be Gaussian of σ=0.3: (a) Noisy image, (b) ROF model [Citation3], (c) NLM model [Citation51], (d) TV+TV2 [Citation19] (e) TGV model [Citation36] and (f) Our model.

Figure 2. The obtained denoised image compared to other classical approaches for the (Pirate image), where the noise is considered to be Gaussian of σ=0.3: (a) Noisy image, (b) ROF model [Citation3], (c) NLM model [Citation51], (d) TV+TV2 [Citation19] (e) TGV model [Citation36] and (f) Our model.

Figure 3. The obtained denoised image compared to other classical approaches for the (Cameraman image), where the noise is considered to be Gaussian of σ=0.5: (a) Noisy image, (b) ROF model [Citation3], (c) NLM model [Citation51], (d) TV+TV2 [Citation19], (e) TGV model [Citation36] and (f) our model.

Figure 3. The obtained denoised image compared to other classical approaches for the (Cameraman image), where the noise is considered to be Gaussian of σ=0.5: (a) Noisy image, (b) ROF model [Citation3], (c) NLM model [Citation51], (d) TV+TV2 [Citation19], (e) TGV model [Citation36] and (f) our model.

Figure 4. The computation of the parameter λ with respect to the iteration for the two images Pirate and Cameraman: (a) The Pirate image and (b) The Cameraman image.

Figure 4. The computation of the parameter λ with respect to the iteration for the two images Pirate and Cameraman: (a) The Pirate image and (b) The Cameraman image.

For the third and fourth experiments, we change the nature of noise which is considered as an impulse one. For that, we change the fidelity term for the compared methods which is now considered as the L1 norm since it success in handling the impulse noise (see [Citation2,Citation54,Citation55]). First, we consider the Cameraman image corrupted by an impulse noise with parameter 0.3. Figure  shows the obtained denoised image using different approaches with L1 norm, we can see once again that the proposed method can efficiently reduce the impulse noise without destroying image features, even though the TGV method reduces the noise better than our approach. For the fourth test, we increase the impulse noise level which is taken with parameter 0.5. Figure  illustrates the restored image using different regularization methods, we can observe that the proposed approach is visually very sharp compared to the TGV result. Also, we plot in Figure  the evolution of the parameter λ with respect to the iterations for the third and fourth examples. It is then observed that the obtained parameter for the Cameraman image is λ=0.81, while for the Penguin image the parameter is λ=0.90. Furthermore, to better perform the proposed denoising method over the other methods, a quantitative evaluation is used. In fact, we provide two metrics: PSNR [Citation56] and the mean structure similarity (SSIM) [Citation57]. The PSNR measures signal strength relative to noise in the image while the SSIM metric gives an indication on the quality of the image related to the known characteristics of human visual system. Tables  and  illustrate the PSNR and SSIM values related to the above simulated tests. We can notice that the proposed model outperforms the compared methods. Moreover, to better compare the quality of reconstructions, we present two tables that indicate the normalized l2 and l1-distances between true image and the restored ones. Tables  and  illustrate the normalized l2 and l1-distances values for the four above simulated tests, respectively. Once again, the proposed approach is always with the lower values, which confirm its robustness.

Figure 5. The obtained denoised image compared to other classical approaches for the Cameraman image, where the noise is considered to be impulse one of parameter 0.3: (a) Noisy image, (b) ROF [Citation3], (c) NLM [Citation51], (d) L1-TV+TV2 [Citation19], (e) L1-TGV [Citation36] and (f) Our model.

Figure 5. The obtained denoised image compared to other classical approaches for the Cameraman image, where the noise is considered to be impulse one of parameter 0.3: (a) Noisy image, (b) ROF [Citation3], (c) NLM [Citation51], (d) L1-TV+TV2 [Citation19], (e) L1-TGV [Citation36] and (f) Our model.

Figure 6. The obtained denoised image compared to other classical approaches for the (Penguin image), where the noise is considered to be impulse one of parameter 0.5: (a) Noisy image, (b) ROF model [Citation3], (c) NLM model [Citation51], (d) L1-TV+TV2 [Citation19], (e) L1-TGV model [Citation36] and (f) Our model.

Figure 6. The obtained denoised image compared to other classical approaches for the (Penguin image), where the noise is considered to be impulse one of parameter 0.5: (a) Noisy image, (b) ROF model [Citation3], (c) NLM model [Citation51], (d) L1-TV+TV2 [Citation19], (e) L1-TGV model [Citation36] and (f) Our model.

Figure 7. The computation of the parameter λ with respect to the iteration for the two images Camerman and Penquin: (a) The Pirate image and (b) The Penguin image.

Figure 7. The computation of the parameter λ with respect to the iteration for the two images Camerman and Penquin: (a) The Pirate image and (b) The Penguin image.

Table 1. The PSNR table.

Table 2. The SSIM table.

Table 3. The normalized l2 distance table.

Table 4. The normalized l1 distance table.

6.2. The space variant case

We now present extensive results to perform the weighted proposed PDE, where the parameter λ is space variant. In fact we are particularly interested to present the results in the space Uad2={λL2(Ω),0<λ_λλ¯<1 a.e. xΩandλ2C0},which is the most used space in the literature (especially λH1(Ω) space), see [Citation27–29,Citation58–60], and also the proposed set of admissible filtering weights Uad1 defined by Uad1={λL1(Ω),0<λ_λλ¯<1 a.e. xΩandTV(λ)C0}.We start by a first experiment where we consider the Lena image which is contaminated by a Gaussian noise with variance σ=0.4. Then, we use the proposed Algorithm 1 to compute the clean image X and the approximate weighted parameter λ using two different choices of the space of admissible weights λ. The restored image and the spatially variant λ (projected in the image grid: λ(X)) using the two sets are displayed in Figure .We can observe that the restored image using Uad1 is little sharp compared to the one restored in Uad2, especially near edges. In particular, the approximate λ for the two cases is higher in homogeneous regions and texture, and weaker near edges where the variation is stronger. The next set of experiments is considered to see the behaviour of the proposed weighted PDE with respect to different noise levels using the two admissible spaces. For that, we consider four images which essentially contain combination of large piecewise constant parts and smooth areas. The restored images using the two sets Uad1 and Uad2 are depicted in Figures  and , respectively. The noisy images are obtained by adding Gaussian noise with variances σ=0.1,0.3,0.4, and 0.5, respectively in the same order from the top to the bottom of Figures and .From the surface representation of λ, we can detect that λ is smooth enough for Uad2 and piecewise constant for the case of Uad1. For the both cases the shapes of λ are related to the one of the obtained image X, even if for Uad1 the shape of λ is almost the same as the picture. In addition, the same remark as for the previous test remains valid since the values of λ are higher near homogeneous areas.

Figure 8. The restored image and the parameter λ using the two admissible sets of the Lena image: (a) Noisy, (b) X using Uad2, (c) λ using Uad2, (d) X using Uad1 and (e) λ using Uad1.

Figure 8. The restored image and the parameter λ using the two admissible sets of the Lena image: (a) Noisy, (b) X using Uad2, (c) λ using Uad2, (d) X using Uad1 and (e) λ using Uad1.

Figure 9. The restored image X and the corresponding weighted parameter λ using the admissible set Uad2. The Baboon image is contaminated by Gaussian noise with σ=0.1, Penguin image is contaminated by σ=0.3, Zebra image is contaminated by σ=0.4, Tiger image is contaminated by σ=0.5. The used parameters for these tests are: (k1,k2)=(35,89), σ=1.3 and ρ=2.7: (a) Noisy, (b) Obtained X, (c) λ and (d) Image of λ.

Figure 9. The restored image X and the corresponding weighted parameter λ using the admissible set Uad2. The Baboon image is contaminated by Gaussian noise with σ=0.1, Penguin image is contaminated by σ=0.3, Zebra image is contaminated by σ=0.4, Tiger image is contaminated by σ=0.5. The used parameters for these tests are: (k1,k2)=(35,89), σ=1.3 and ρ=2.7: (a) Noisy, (b) Obtained X, (c) λ and (d) Image of λ.

Figure 10. The restored image X and the corresponding weighted parameter λ using the admissible set Uad1. the Baboon image is contaminated by Gaussian noise with σ=0.1, Penguin image is contaminated by σ=0.3, Zebra image is contaminated by σ=0.4, Tiger image is contaminated by σ=0.5. The used parameters for these tests are: (k1,k2)=(35,80), σ=1.7 and ρ=2.3: (a) Noisy, (b) Obtained X, (c) Obtained λ and (d) Image of λ.

Figure 10. The restored image X and the corresponding weighted parameter λ using the admissible set Uad1. the Baboon image is contaminated by Gaussian noise with σ=0.1, Penguin image is contaminated by σ=0.3, Zebra image is contaminated by σ=0.4, Tiger image is contaminated by σ=0.5. The used parameters for these tests are: (k1,k2)=(35,80), σ=1.7 and ρ=2.3: (a) Noisy, (b) Obtained X, (c) Obtained λ and (d) Image of λ.

The following four tests concern the comparison between the proposed weighted PDE and the spatially dependent regularization parameters for TV image denoising (SDRTV) [Citation27]. For a fair comparison, we consider the same conditions for our approach as for SDRTV one (where λH1(Ω)). For that, we set λUad2 for our method and we consider two comparative results. In the two experiments, we treat the denoising problem with brain scan images. The first set consists of image of 256×256 pixels and Gaussian noise with zero mean and variance σ=0.3. The second one consists of image of 512×512 pixels and Gaussian noise with zero mean and variance σ=0.4. The used parameters for SDRTV method are γ=40,1013, μ=1012,β=1010 and h=0.1. The computed semismooth Newton method, on the whole domain, converges when kmax = 40 iterations. Note that the compared results are done without domain decomposition. For our method, the used parameters are (k1,k2)=(55,40), σ=1.8 and ρ=2.1. The corresponding results are shown in Figures  and , respectively. By a visual comparison, we can see that the proposed PDE yields significant improvement compared to the SDRTV method. In addition, from the surface representation of λ for the two methods, we can see that λ is continuous and its shape is related to the one of the original image (especially in homogeneous areas).

Figure 11. Comparison between the obtained clean image X and the SDRTV method with the respective computation of the spatially dependent parameter λ of the Head image: (a) Original, (b) Noisy, (c) λ initial, (d) SDRTV, (e) Obtained λ, (f) Image of λ, (g) Proposed PDE, (h) Obtained λ and (i) Image of λ.

Figure 11. Comparison between the obtained clean image X and the SDRTV method with the respective computation of the spatially dependent parameter λ of the Head image: (a) Original, (b) Noisy, (c) λ initial, (d) SDRTV, (e) Obtained λ, (f) Image of λ, (g) Proposed PDE, (h) Obtained λ and (i) Image of λ.

Figure 12. Comparison between the obtained clean image and the SDRTV method and the respective computation of the spatially dependent parameter λ of the Brain image: (a) Original, (b) Noisy, (c) λ initial, (d) SDRTV, (e) Obtained λ, (f) Image of λ, (g) SDRTV, (h) Obtained λ and (i) Image of λ.

Figure 12. Comparison between the obtained clean image and the SDRTV method and the respective computation of the spatially dependent parameter λ of the Brain image: (a) Original, (b) Noisy, (c) λ initial, (d) SDRTV, (e) Obtained λ, (f) Image of λ, (g) SDRTV, (h) Obtained λ and (i) Image of λ.

We now consider the proposed admissible set such as λUad1 for our method and we propose two experiment results where we compare our PDE with the SDRTV approach (λH1(Ω)). We keep the same parameters as for the previous tests. The noise is considered to be Gaussian one with variances σ=0.3 and σ=0.5, respectively. The recovered images are depicted in Figures  and , respectively. Once again, our method outperforms the SDRTV by a visual comparison. In addition the approximate λ tends to capture more information from the image, which can be interpreted by the less regularity of the proposed set of admissible λ.

Figure 13. Comparison between the obtained clean image X and the SDRTV method with the respective computation of the spatially dependent parameter λ of the Dolphins image: (a) Original, (b) Noisy, (c) λ initial, (d) SDRTV, (e) Obtained λ, (f) Image of λ, (g) Obtained X, (h) Obtained λ and (i) Image of λ.

Figure 13. Comparison between the obtained clean image X and the SDRTV method with the respective computation of the spatially dependent parameter λ of the Dolphins image: (a) Original, (b) Noisy, (c) λ initial, (d) SDRTV, (e) Obtained λ, (f) Image of λ, (g) Obtained X, (h) Obtained λ and (i) Image of λ.

Figure 14. Comparison between the obtained clean image X and the SDRTV method with the respective computation of the spatially-dependent parameter λ of the Plane image: (a) Original, (b) Noisy, (c) λ initial, (d) SDRTV, (e) Obtained λ, (f) Image of λ, (g) Obtained X, (h) Obtained λ and (i) Image of λ.

Figure 14. Comparison between the obtained clean image X and the SDRTV method with the respective computation of the spatially-dependent parameter λ of the Plane image: (a) Original, (b) Noisy, (c) λ initial, (d) SDRTV, (e) Obtained λ, (f) Image of λ, (g) Obtained X, (h) Obtained λ and (i) Image of λ.
We end this part by a final example where we give a comparison between the scalar approximate λ, the weighted spatial approximate λ in the respective spaces Uad1 and Uad2. We consider then the image of Fishes, which is a textured challenging example. Then, we add a Gaussian noise with high variance σ=0.5. The restored image using different λ approximations is presented in Figure . In view of the subplots in Figure , the obtained spatially variant regularization parameter shares similar patterns to the ones obtained in the previous tests. Particularly, the restored image X using the admissible set Uad1 is visually even better, especially near sharp edges. In addition, the spatially variant result leads to a reduction of the staircasing effect better than the scalar result.

6.3. Robustness of the proposed PDE

In this part, we use four simulated images to perform the choice of the proposed PDE-constrained with the space variant choice of λ. To show the efficiency of the proposed PDE in recovering essential image features, we use some comparisons to other competitive denoising PDEs. In fact, the obtained results are compared to the Fourth-order partial differential equations for image enhancement (FOPDE) [Citation61], Adaptive fourth-order partial differential equation filter for image denoising (AEFD) [Citation62] and the adaptive fourth-order partial differential equation for image denoising (AFOD) [Citation63]. The four tests are done using different levels of Gaussian noise, with σ=30,40,50 and 60, respectively. In Figure , we present the restored image of the four noisy images using the compared enhancement PDEs and our. We can see obviously the robustness of the proposed PDE in avoiding different artefacts compared with the other ones. The used parameters are tuned with respect to the best obtained PSNR. For example the used parameters for the Tiger image are: λ=0.09, Δt=0.07, σ=0.5, K=0.001 and 50 iterations for the AEFD approach. μ=0.09, Δt=0.05, k1=0.05, k2=0.05, γ=106 and 100 iterations for the AFOD approach. β=104, Δt=0.1, σ1=0.001, σ2=0.8,α1=0.01, α2=44 and 120 iterations for the FOPDE method. Note that the used parameters of our approach are: (k1,k2)=(33,42), σ=1.54, ρ=3.1, β=1 and 50 iterations number.

Figure 15. Comparison between the obtained clean image X and the respective computation of the spatially dependent parameter λ for the scalar case and for it two set of admissible values Uad1 and Uad2 for the Fishes image: (a) Original, (b) Noisy, (c) scalar λ, (d) weighted λUad2 and (e) weighted λUad1.

Figure 15. Comparison between the obtained clean image X and the respective computation of the spatially dependent parameter λ for the scalar case and for it two set of admissible values Uad1 and Uad2 for the Fishes image: (a) Original, (b) Noisy, (c) scalar λ, (d) weighted λ∈Uad2 and (e) weighted λ∈Uad1.

Figure 16. The obtained denoised image compared to other PDE approaches with respect to both quality measures PSNR and SSIM. First row: noisy images. Second row: FOPDE. Third row: AEFD. Fourth row: AFOD. Fifth row: Our approach. (a) PSNR = 19.17, SSIM = 0.327. (b) PSNR = 15.40, SSIM = 0.377. (c) PSNR = 15.12, SSIM = 0.310. (d) PSNR = 14.08, SSIM = 0.286. (e) PSNR = 23.79, SSIM = 0.539. (f) PSNR = 22.61, SSIM = 0.588. (g) PSNR = 20.66, SSIM = 0.548. (h) PSNR = 20.41, SSIM = 0.422, (i) PSNR = 26.19, SSIM = 0.719, (j) PSNR = 23.95, SSIM = 0.650, (k) PSNR = 23.62, SSIM = 0.647. (l) PSNR = 28.48, SSIM = 0.747. (m) PSNR = 26.94, SSIM = 0.746, (n) PSNR = 25.62, SSIM = 0.741, (o) PSNR = 22.81, SSIM = 0.627, (p) PSNR = 24.96, SSIM = 0.597, (q) PSNR = 28.14, SSIM = 0.826, (r) PSNR = 27.65, SSIM = 0.821, (s) PSNR = 26.03, SSIM = 0.765 and (t) PSNR = 31.91, SSIM = 0.872.

Figure 16. The obtained denoised image compared to other PDE approaches with respect to both quality measures PSNR and SSIM. First row: noisy images. Second row: FOPDE. Third row: AEFD. Fourth row: AFOD. Fifth row: Our approach. (a) PSNR = 19.17, SSIM = 0.327. (b) PSNR = 15.40, SSIM = 0.377. (c) PSNR = 15.12, SSIM = 0.310. (d) PSNR = 14.08, SSIM = 0.286. (e) PSNR = 23.79, SSIM = 0.539. (f) PSNR = 22.61, SSIM = 0.588. (g) PSNR = 20.66, SSIM = 0.548. (h) PSNR = 20.41, SSIM = 0.422, (i) PSNR = 26.19, SSIM = 0.719, (j) PSNR = 23.95, SSIM = 0.650, (k) PSNR = 23.62, SSIM = 0.647. (l) PSNR = 28.48, SSIM = 0.747. (m) PSNR = 26.94, SSIM = 0.746, (n) PSNR = 25.62, SSIM = 0.741, (o) PSNR = 22.81, SSIM = 0.627, (p) PSNR = 24.96, SSIM = 0.597, (q) PSNR = 28.14, SSIM = 0.826, (r) PSNR = 27.65, SSIM = 0.821, (s) PSNR = 26.03, SSIM = 0.765 and (t) PSNR = 31.91, SSIM = 0.872.

7. Conclusion

In this paper, we have introduced a fourth-order PDE-constrained optimization model to treat the denoising task. This PDE is elaborated using a convex combination between two operators controlled by a weighting parameter, which is the solution of the optimal problem. This PDE takes the advantage of the Perona–Malik equation with much regularity and the efficiency of a nonlinear filter to restore tiny and sharp edges. The existence and uniqueness of the proposed PDE-constrained is proved using the Schauder fixed point theorem and the well-posedness of the optimization model is established using the control theory in a suitable space. The results part demonstrates, visually and quantitatively the performance and the contribution of this new PDE over the other denoising approaches. In summary, the proposed approach gives promising analytic strategy for determining an optimal solution for the denoised image and also the parameter λ. As further developments of this method, it may be generalized for other operators and different regularization functional. Another very important generalization of the proposed model is to move from a simple denoising model to more general inverse problems involving linear and nonlinear operators.

Acknowledgments

We are very grateful to the anonymous referees for the remarkable corrections and useful suggestions that have much improved this paper.

Disclosure statement

No potential conflict of interest was reported by the authors.

References

  • Kirsch A. An introduction to the mathematical theory of inverse problems. Vol. 120. Germany: Springer Science & Business Media; 2011.
  • Chan TF, Esedoglu S. Aspects of total variation regularized l 1 function approximation. SIAM J Appl Math. 2005;65(5):1817–1837.
  • Rudin LI, Osher S, Fatemi E. Nonlinear total variation based noise removal algorithms. Phys D Nonlinear Phenom. 1992;60(1-4):259–268.
  • Hintermüller M, Holler M, Papafitsoros K. A function space framework for structural total variation regularization with applications in inverse problems. Inverse Probl. 2018;34(6):064002.
  • Liu J, Ni G, Yan S. Alternating method based on framelet l 0-norm and tv regularization for image restoration. Inverse Probl Sci Eng. 2019;27(6):790–807.
  • Laghrib A, Ezzaki M, El Rhabi M, et al. Simultaneous deconvolution and denoising using a second order variational approach applied to image super resolution. Comput Vis Image Underst. 2018;168:50–63.
  • Chantas G, Galatsanos NP, Molina R, et al. Variational bayesian image restoration with a product of spatially weighted total variation image priors. IEEE Trans Image Process. 2010;19(2):351–362.
  • Rodríguez P, Wohlberg B. Efficient minimization method for a generalized total variation functional. IEEE Trans Image Process. 2009;18(2):322–332.
  • Frick K, Marnitz P, Munk A. Statistical multiresolution estimation for variational imaging: with an application in poisson-biophotonics. J Math Imaging Vis. 2013;46(3):370–387.
  • Dong Y, Hintermüller M, Rincon-Camacho MM. Automated regularization parameter selection in multi-scale total variation models for image restoration. J Math Imaging Vis. 2011;40(1):82–104.
  • Bertalmío M, Caselles V, Rougé B, et al. Tv based image restoration with local constraints. J Sci Comput. 2003;19(1–3):95–122.
  • Almansa A, Ballester C, Caselles V, et al. A tv based restoration model with local constraints. J Sci Comput. 2008;34(3):209–236.
  • Strong DM, Aujol J-F, Chan TF. Scale recognition, regularization parameter selection, and Meyer's g norm in total variation regularization. Multiscale Model Simulat. 2006;5(1):273–303.
  • Gilboa G, Sochen NA, Zeevi YY. Estimation of optimal pde-based denoising in the snr sense. IEEE Trans Image Proces. 2006;15(8):2269–2280.
  • Vogel CR. Computational methods for inverse problems. Vol. 23. US: SIAM; 2002.
  • De los Reyes JC, Schönlieb C-B. Image denoising: learning the noise model via nonsmooth pde-constrained optimization. Inverse Problems Imaging. 2013;7(4):1183–1214.
  • Chan T, Marquina A, Mulet P. High-order total variation-based image restoration. SIAM J Sci Comput. 2000;22(2):503–516.
  • Osher S, Solé A, Vese L. Image decomposition and restoration using total variation minimization and the h. Multiscale Model Simul. 2003;1(3):349–370.
  • Papafitsoros K, Schönlieb C-B. A combined first and second order variational approach for image reconstruction. J Math Imaging Vis. 2014;48(2):308–338.
  • Chen Y, Ranftl R, Pock T. Insights into analysis operator learning: from patch-based sparse models to higher order mrfs. IEEE Trans Image Process. 2014;23(3):1060–1072.
  • Haber E, Tenorio L. Learning regularization functionals–a supervised training approach. Inverse Probl. 2003;19(3):611.
  • De los Reyes JC, Schönlieb C-B, Valkonen T. The structure of optimal parameters for image restoration problems. J Math Anal Appl. 2016;434(1):464–500.
  • De los Reyes JC, Schönlieb C-B, Valkonen T. Bilevel parameter learning for higher-order total variation regularisation models. J Math Imaging Vis. 2017;57(1):1–25.
  • Chung C, De los Reyes JC, Schönlieb C-B. Learning optimal spatially-dependent regularization parameters in total variation image restoration. arXiv preprint arXiv:1603.09155.
  • Kunisch K, Pock T. A bilevel optimization approach for parameter learning in variational models. SIAM J Imaging Sci. 2013;6(2):938–983.
  • Calatroni L, Cao C, De Los Reyes JC, et al. Bilevel approaches for learning of variational imaging models. Variat Methods Imaging Geom Control. 2017;18(252):2.
  • Van Chung C, De los Reyes J, Schönlieb C. Learning optimal spatially-dependent regularization parameters in total variation image denoising. Inverse Probl. 2017;33(7):074005.
  • Hintermüller M, Rautenberg CN. Optimal selection of the regularization function in a weighted total variation model. part i: modelling and theory. J Math Imaging Vis. 2017;59(3):498–514.
  • Hintermüller M, Rautenberg CN, Wu T, et al. Optimal selection of the regularization function in a weighted total variation model. part ii: algorithm, its analysis and numerical tests. J Math Imaging Vision. 2017;59(3):515–533.
  • Perona P, Shiota T, Malik J. Anisotropic diffusion. In: Geometry-driven diffusion in computer vision. Springer; 1994. p. 73–92.
  • Weickert J. Coherence-enhancing diffusion filtering. Journal of Computer Vision¡/DIFdel¿Int J Comput Vis. 1999;31(2-3):111–127.
  • Weickert J, Scharr H. A scheme for coherence-enhancing diffusion filtering with optimized rotation invariance. J Vis Commun Image Represent. 2002;13(1-2):103–118.
  • Burgeth B, Didas S, Weickert J. A general structure tensor concept and coherence-enhancing diffusion filtering for matrix fields. In: Visualization and processing of tensor fields. Springer; 2009. p. 305–323.
  • Elad M. On the origin of the bilateral filter and ways to improve it. Image Processing¡/DIFdel¿IEEE Trans Image Process. 2002;11(10):1141–1151.
  • El Mourabit I, El Rhabi M, Hakim A, et al. A new denoising model for multi-frame super-resolution image reconstruction. Signal Processing. 2017;132:51–65.
  • Bredies K, Kunisch K, Pock T. Total generalized variation. SIAM J Imaging Sci. 2010;3(3):492–526.
  • Valkonen T, Bredies K, Knoll F. Total generalized variation in diffusion tensor imaging. SIAM J Imaging Sci. 2013;6(1):487–525.
  • Chambolle A, Pock T. A first-order primal-dual algorithm for convex problems with applications to imaging. J Math Imaging Vis. 2011;40(1):120–145.
  • Zhang X, Burger M, Osher S. A unified primal-dual algorithm framework based on bregman iteration. J Sci Comput. 2011;46(1):20–46.
  • Laghrib A, Hakim A, Raghay S. An iterative image super-resolution approach based on bregman distance. Signal Proces Image Commun. 2017;58:24–34.
  • Afraites L, Hadri A, Laghrib A. A denoising model adapted for impulse and gaussian noises using a constrained-pde. Inverse Probl. 2020;36(2):025006.
  • Gilbarg D, Trudinger NS. Elliptic partial differential equations of second order. Springer; 2015.
  • Catté F, Lions P-L, Morel J-M, et al. Image selective smoothing and edge detection by nonlinear diffusion. SIAM J Numer Anal. 1992;29(1):182–193.
  • Brezis H. Analyse fonctionnelle. Masson: Paris; 1983.
  • Dautray R, Lions J. Mathematical analysis and numerical methods for science and technology: evolution problems I. US: Springer; 1992.
  • Zeidler E. Nonlinear functional analysis and its applications: III: variational methods and optimization. US: Springer Science & Business Media; 2013.
  • Aubin JP. Un théorème de compacité. Acad Sci Paris. 1963;256:5042–5044.
  • Simon J. Compact sets in the space lp(0,t;b). Ann. Mat. Pura Appl. 1987;146:65–96.
  • Clason C, Valkonen T. Primal-dual extragradient methods for nonlinear nonsmooth pde-constrained optimization. SIAM J Optim. 2017;27(3):1314–1339.
  • Wu C, Tai X-C. Augmented lagrangian method, dual methods, and split Bregman iteration for rof, vectorial tv, and high order models. SIAM J Imaging Sci. 2010;3(3):300–339.
  • Buades A, Coll B, Morel J-M. Non-local means denoising. Image Process On Line. 2011;1:208–212.
  • Maleki A, Narayan M, Baraniuk RG. Anisotropic nonlocal means denoising. Appl Comput Harmon Anal. 2013;35(3):452–482.
  • Papafitsoros K, Schoenlieb CB, Sengul B. Combined first and second order total variation inpainting using split Bregman. Image Process On Line. 2013;3:112–136.
  • Duval V, Aujol J-F, Gousseau Y. The tvl1 model: a geometric point of view. Multiscale Model Simul. 2009;8(1):154–189.
  • Nikolova M. A variational approach to remove outliers and impulse noise. J Math Imaging Vis. 2004;20(1-2):99–120.
  • De Boer JF, Cense B, Park BH, et al. Improved signal-to-noise ratio in spectral-domain compared with time-domain optical coherence tomography. Opt Lett. 2003;28(21):2067–2069.
  • Wang Z, Simoncelli EP, Bovik AC. Multiscale structural similarity for image quality assessment. In: The thirty-seventh asilomar conference on signals, systems & computers. Vol. 2. IEEE; 2003, p. 1398–1402.
  • Bredies K, Dong Y, Hintermüller M. Spatially dependent regularization parameter selection in total generalized variation models for image restoration. Int J Comput Math. 2013;90(1):109–123.
  • Hintermüller M, Papafitsoros K. Generating structured nonsmooth priors and associated primal-dual methods. In: Handbook of numerical analysis. Vol. 20. Elsevier; 2019. p. 437–502.
  • Hintermüller M, Papafitsoros K, Rautenberg CN, et al. Dualization and automatic distributed parameter selection of total generalized variation via bilevel optimization. arXiv preprint arXiv:2002.05614.
  • Yi D, Lee S. Fourth-order partial differential equations for image enhancement. Appl Math Comput. 2006;175(1):430–440.
  • Liu X, Huang L, Guo Z. Adaptive fourth-order partial differential equation filter for image denoising. Appl Math Lett. 2011;24(8):1282–1288.
  • Zhang X, Ye W. An adaptive fourth-order partial differential equation for image denoising. Comput Math Appl. 2017;74(10):2529–2545.
  • Casas E, Fernandez L. Distributed control of systems governed by a general class of quasilinear elliptic equations. J Diff Equ.
  • Ciarlet PG. The finite element method for elliptic problems. Vol. 40. US: Siam; 2002.
  • Eymard R, Herbin R, Linke A, et al. Convergence of a finite volume scheme for the biharmonic problem.

Appendices

Appendix 1

The next propositions show the differentiability of the solution operator S. This result leads us to get an expression of the operator S(λ) using the proper adjoint state.

Proposition A.1

Let S:UadL2(Ω) be the solution of Equation (Equation2) at t = T associated to each parameter λ. The operator S is Gâteaux differentiable and its derivative at λ, in direction h, is given by the unique solution X1 of the following linearized equation: (A1) X1t+2.λD(Jρσ(X))2X1.(1λ)D(Jρσ(X))X1+2.λD(Jρσ(X))X12X.(1λ)D(Jρσ(X))X1X=2.hD(Jρσ(X))2X.hD(Jρσ(X))X in ]0,T[×Ω,D(Jρσ(X))X1=0 on ]0,T[×Ω,X1(t,x)=0, on (0,T)×Ω,X1(0,x)=0, in Ω,(A1) where D(Jρσ(X))=D(Jρσ(X))(Jρσ(X)) is the derivative of D(Jρσ(X)) with respect to state X.

Proof.

We decompose the proof in tree steps:

Step 1 Let Xϵ and X be the unique solutions to (Equation2) corresponding to λ+ϵh and λ, respectively for hUadL(Ω)BV(Ω) and small ϵ. Let's show the priori estimations of the difference of the two solutions Xϵ and X. Taking the difference between the equations of Xϵ and X, we have: (A2) XϵtXt,ϕ+Ωλ(D(Jρσ(Xϵ))2XϵD(Jρσ(X))2X)2ϕdx+Ω(1λ)(D(Jρσ(Xϵ))XϵD(Jρσ(X))X)ϕdx=ϵΩhD(Jρσ(Xϵ))(2Xϵ2ϕXϵϕ)dx,(A2) we can rewrite (EquationA2) as follows: (A3) (XϵX)t,ϕ+ΩλD(Jρσ(Xϵ))(2(XϵX))2ϕdx+Ω(1λ)D(Jρσ(Xϵ))((XϵX))ϕdx+Ωλ(D(Jρσ(Xϵ))D(Jρσ(X)))2X2ϕ+Ω(1λ)(D(Jρσ(Xϵ))D(Jρσ(X)))Xϕdx=ϵΩhD(Jρσ(Xϵ))(2Xϵ2ϕXϵϕ)dx.(A3) (A4) 2(XϵX)L2(0,T;L2(Ω))CϵhL(Ω),(A4) (A5) (XϵX)L2(0,T;L2(Ω))CϵhL(Ω),(A5) (A6) XϵXL(0,T;L02(Ω))CϵhL(Ω),(A6) and (A7) (XϵX)tL2(0,T;H2(Ω))CϵhL(Ω).(A7) Step 2 In the second step, we consider the sequence {Xϵ1}ϵ>0, with Xϵ1:=XϵXϵ and we prove the weak convergence of Xϵ1:=XϵXϵ to X1 solution of Equation (EquationA1).

We have the corresponding Equation (EquationA3) at Xϵ1 is given as follows: (A8) Xϵ1t,ϕ+ΩλD(Jρσ(Xϵ))2Xϵ12ϕdx+ΩλD(Jρσ(Xϵ))D(Jρσ(X))ϵ2X2ϕ+Ω(1λ)D(Jρσ(Xϵ))Xϵ1ϕdx+Ω(1λ)D(Jρσ(Xϵ))D(Jρσ(X))ϵXϕdx=ΩhD(Jρσ(Xϵ))(2Xϵ2ϕXϵϕ)dx.(A8) Using the mean value theorem in integral form we get that: (A9) Xϵ1t,ϕ+ΩλD(Jρσ(Xϵ))2Xϵ12ϕdx+ΩλD(Jρσ(Wϵ))Xϵ12X2ϕ+Ω(1λ)D(Jρσ(Xϵ))Xϵ1ϕdx+Ω(1λ)D(Jρσ(Wϵ))Xϵ1Xϕdx=ΩhD(Jρσ(Xϵ))(2Xϵ2ϕXϵϕ)dx.(A9) where Wϵ=X+ρϵ(XϵX) with 0ρϵ1.

Using the estimations (EquationA4) and (EquationA7) Aubin–Lions and Aubin–Simon lemmas [Citation45–48], we can extract a subsequence denoted again (Xϵ1) such that Xϵ1X1 strongly in L2(0,T;H01(Ω)) and Xϵ1X1 strongly in L(0,T;L2(Ω)) as ϵ0. Using the same techniques of convergence we can extract a subsequence denoted again (Xϵ) such that XϵX strongly in L2(0,T;H01(Ω)) and XϵX strongly in L(0,T;L2(Ω)), and thanks to the smoothness of D, and form the definition we have D is continuous. We can pass to the limit in (EquationA9) to obtain the following formula: (A10) X1t,ϕ+ΩλD(Jρσ(X))2X12ϕdx+ΩλD(Jρσ(X)X12X2ϕ+Ω(1λ)D(Jρσ(X))X1ϕdx+Ω(1λ)D(Jρσ(X)X1Xϕdx=ΩhD(Jρσ(X))(2X2ϕXϕ)dx.(A10) Consequently, X1 corresponds to the solution of the linearised equation (EquationA1) with (A11) 2X1L2(0,T;L2(Ω))ChL(Ω),(A11) (A12) X1L2(0,T;L2(Ω))ChL(Ω),(A12) (A13) X1L(0,T;L02(Ω))ChL(Ω),(A13) and (A14) X1tL2(0,T;H2(Ω))ChL(Ω).(A14) Step 3 Now, we will prove that the sequence Xϵ1 converges strongly in L2(0,T;H02(Ω)) to X1. As a consequence of the previous step, it suffices to prove that: 2Xϵ12X1in (L2(Ω))2Let consider two Matrix: Mϵ=(λ+ϵh)D(Jρσ(Xϵ))andM=D(Jρσ(X))Mϵ and M are symmetric and positive definite. Using Cholesky decomposition we obtain lower triangular matrices Lϵ and L such that Mϵ=LϵLϵTandM=LLT.Applying the same convergence as in step 2 in the weak formulation (EquationA9) with the function test is Xϵ1 and integrating by parts we obtain (A15) limϵ00TΩMϵ2Xϵ12Xϵ1dxdt=12X1(.,T)L2(Ω)20TΩ(1λ)D(Jρσ(X))X12X2X1dxdt0TΩ(1λ)D(Jρσ(X))X1X1dxdt0TΩ(1λ)D(Jρσ(X))X1XX1dxdt+0TΩhD(Jρσ(X))(2X2X1XX1)dxdt=0TΩM2X12X1dxdt0TΩMϵ2Xϵ12Xϵ1dxdt0TΩM2X12X1dxdtas ϵ0,(A15) Using the above equation and (EquationA10), we prove by the same argumentation as in ([Citation64], Theorem 3.1, Step 3) that which implies that 2Xϵ12X1 in L2(0,T;L2(Ω)).

We can find by using the same techniques that Xϵ1tX1t in L2(0,T;H2(Ω)). This complete the proof.

Based on the differentiability properties of the solution operator, we give in next proposition the expression of adjoint derivative operator S(λ)(.).

Proposition A.2

We have: (A16) S(λ)W=TD(Jρσ(X))2X2Pdt+0TD(Jρσ(X))XPdt,(A16) where P is the solution of the following adjoint problem: (A17) Pt+2.λD(Jρσ(X))2P.(1λ)D(Jρσ(X))P+λD(Jρσ(X))2X2P+(1λ)D(Jρσ(X))XP=0in]0,T[×Ω,D(Jρσ(X))P,ν=0on]0,T[×Ω,P(t,x)=0, on ,(0,T)×Ω,P(T,x)=W(x), in Ω,(A17)

Proof.

To prove the result (Equation74), we take ϕL2(0,T;H02(Ω)) as function test in the variational form of system (EquationA1) and (Equation75), which give: (A18) 0TX1t,ϕ+0TΩλD(Jρσ(X))2X12ϕ+0TΩ(1λ)D(Jρσ(X))X1ϕ+0TΩλD(Jρσ(X))X12X2ϕ+0TΩ(1λ)D(Jρσ(X))X1Xϕ=0TΩhD(Jρσ(X))2Xϕ+0TΩhD(Jρσ(X))Xϕ(A18) and (A19) 0TPt,ϕ+0TΩλD(Jρσ(X))2P2ϕ+Ω(1λ)D(Jρσ(X))Pϕ+0TΩλD(Jρσ(X))2X2Pϕ+0TΩ(1λ)D(Jρσ(X))XPϕ=0(A19) Taking ϕ=P in Equation (EquationA18) and ϕ=X1 in Equation (EquationA19) and use the formula 0TX1t,P=0TPt,X1+ΩX1(x,T)P(x,T)dx,it immediately shows that ΩX1(x,T)P(x,T)dx=0TΩhD(Jρσ(X))2X2Pdt+0TΩhD(Jρσ(X))XPdxdtTo finish this the proof, we use the proposed notation X1(T)=K(λ)h=S(λ)h and P(T)=W then, we have S(λ)h,W=0T(D(Jρσ(X))2X2PD(Jρσ(X))XP)dt,h.which completes the proof.

Appendix 2

We introduce the functional space H02(Ω), which is defined as the closure of Cc(Ω) in H2(Ω). Thanks to the Lipschitz regularity of the boundary, which leads to the definition H02(Ω)=uH01(Ω)H2(Ω) such that un=0 a.e. on Ω.The space H02(Ω) equipped by the norm ΔuL2(Ω), which is equivalent to the uH2(Ω) for every uH02(Ω) (see [Citation65,Citation66]). Indeed, the Poincaré inequality uH01(Ω),uL2(Ω)CuL2(Ω)2and uH02(Ω),ΩuΔudx=Ωuudx,leads to uH02(Ω),uL2(Ω)2CΔuL2(Ω).Besides, the following equality which is an immediate consequence of two integrations by parts ϕCc(Ω),Ω(Δϕ)2dx=i=12j=12Ω2ϕ2xi2ϕ2xjdx=i=12j=12Ω(2ϕxixj)2dx.Now we give the definition of some useful operators. We have div2(A)=2A12x1x2+2A21x2x1+i=122Aii2xi,ACc(Ω,R2×2)and 2f=2f2x12fx1x22fx2x12f2x2,fCc(Ω).

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.