Home Penalty-based smoothness conditions in convex variational regularization
Article
Licensed
Unlicensed Requires Authentication

Penalty-based smoothness conditions in convex variational regularization

  • Bernd Hofmann ORCID logo EMAIL logo , Stefan Kindermann and Peter Mathé
Published/Copyright: March 8, 2019

Abstract

The authors study Tikhonov regularization of linear ill-posed problems with a general convex penalty defined on a Banach space. It is well known that the error analysis requires smoothness assumptions. Here such assumptions are given in form of inequalities involving only the family of noise-free minimizers along the regularization parameter and the (unknown) penalty-minimizing solution. These inequalities control, respectively, the defect of the penalty, or likewise, the defect of the whole Tikhonov functional. The main results provide error bounds for a Bregman distance, which split into two summands: the first smoothness-dependent term does not depend on the noise level, whereas the second term includes the noise level. This resembles the situation of standard quadratic Tikhonov regularization in Hilbert spaces. It is shown that variational inequalities, as these were studied recently, imply the validity of the assumptions made here. Several examples highlight the results in specific applications.

MSC 2010: 47A52; 65J20

Funding source: Austrian Science Fund

Award Identifier / Grant number: P 30157-N31

Award Identifier / Grant number: HO 1454/12-1

Funding statement: The research of the first author was supported by Deutsche Forschungsgemeinschaft (DFG-grant HO 1454/12-1). The research of the second author was supported by the Austrian Science Fund (FWF) project P 30157-N31.

A Proofs

Let us define the noisy and exact residuals, and the noise term as

pαδ:=Axαδ-yδ,pα:=Axα-y=A(xα-x),Δ:=yδ-y.

Notice that all quantities pα, pαδ as well as Δ belong to the Hilbert space H. The subsequent analysis will be based on the optimality conditions (recall our convention on the choice of ξαδJ(xαδ) and ξαJ(xα))

(A.1)Axαδ-yδ,Aw+αξαδ,w=0for all wX,
(A.2)Axα-Ax,Aw+αξα,w=0for all wX.

In particular, the optimality conditions lead to the following formulas, by

  1. subtracting (A.1) from (A.2) using w=xα-x,

  2. using (A.1) with w=xα-xαδ,

  3. using (A.2) with w=xα-x,

respectively:

(A.3)ξα-ξαδ,x-xα=-1αpαδ-pα,pα,
(A.4)-ξαδ,xα-xαδ=1αpαδ,pα-pαδ-Δ),
(A.5)-ξα,x-xα=-1αpα2.

The following bounds will be the key for proving Theorem 2.8.

Lemma A.1.

Under Assumption 2.3 we have

(A.6)Bξα(xα;x)Ψ(α)-12αpα2,
Bξαδ(xαδ;xα)δ22α-12αpα2+1αpαδ,pα.

Proof.

Using the optimality condition (A.5), we find that

Bξα(xα;x)+12αpα2=J(x)-J(xα)-ξα,x-xα+12αpα2
=J(x)-J(xα)-12αpα2
=1α(Tα(x;y)-Tα(xα;y))Ψ(α),

which proves the first assertion.

For proving the second assertion we use the definition of Bξαδ(xαδ;xα) and (A.4) to find

(A.7)Bξαδ(xαδ;xα)=J(xα)-J(xαδ)+1αpαδ,pα-pαδ-Δ.

The minimizing property of xα also yields

J(xα)-J(xαδ)12α[Axαδ-y2-Axα-y2]=12α[pαδ+Δ2-pα2].

We rewrite

pαδ,pα-pαδ-Δ=-pαδ2+pαδ,pα-Δ.

Using this and plugging the above estimate into (A.7) gives

Bξαδ(xαδ;xα)12α[pαδ+Δ2-pα2]-1αpαδ2+1αpαδ,pα-Δ
=12αΔ2-12αpα2-12αpαδ2+1αpαδ,pα
δ22α-12αpα2+1αpαδ,pα,

completing the proof of the second assertion and of the lemma. ∎

We are now in a position to give detailed proofs of the results in Section 2.

Proof of Proposition 2.4.

From (A.6) we know that, for all α>0,

Axα-y22α=pα22αΨ(α).

Therefore, Assumption 2.3 implies that

J(x)-J(xα)=J(x)-J(xα)-Axα-y22α+pα22αΨ(α)+pα22α2Ψ(α),

which completes the proof. ∎

Proof of Proposition 2.6.

First, if infxXJ(x)=J(x), then

αJ(x)=Tα(x;Ax)Tα(xα;Ax)αJ(xα)αJ(x),

and for all α>0 we have J(xα)=J(x). This allows us to prove the assertion in the first (singular) case. Otherwise, assume to the contrary that there is an index function Ψ, and for a decreasing sequence of regularization parameters (αk)k with limkαk=0 the limit condition

limkΨ(αk)αk=0

holds. Consequently, we find from (2.5) that limk1αkAxαk-y=0, with y=Ax.

Due to the optimality condition (A.2) for xα, we have that

A*(Axαk-y)+αkξαk=0for some ξαkJ(xαk)X*.

This yields

ξαk=-1αkA*(Axαk-y).

Since A*:HX* is a bounded linear operator, we get

ξαkX*1αkA*(H,X*)Axαk-y0as k.

Since ξαkJ(xαk) and after taking the limit, we find for all zX that

J(z)lim supk{J(xαk)+ξαk,z-xαk}=lim supkJ(xαk)=J(x),

where we used Assumption 2.2 and limkΨ(αk)=0. Thus, we conclude that infxXJ(x)=J(x). This contradicts the assumption, and hence the function Ψ cannot decrease to zero super-linearly as α0. ∎

Proof of Theorem 2.8.

Here we recall the three-point identity (see, e.g., [29]). For u,v,wX and ξJ(w), ηJ(v), we have that

Bξ(w;u)=Bη(v;u)+Bξ(w;v)+η-ξ,u-v,

and this specifies with u:=x, v:=xα, and w:=xαδ to

(A.8)Bξαδ(xαδ;x)=Bξα(xα;x)+Bξαδ(xαδ;xα)+ξα-ξαδ,x-xα.

Inserting (A.3) into (A.8) gives

Bξαδ(xαδ;x)=Bξα(xα;x)+Bξαδ(xαδ;xα)-1αpαδ-pα,pα.

An application of the bounds in Lemma A.1 provides us with the estimate

Bξαδ(xαδ;x)Ψ(α)+δ22α-1αpα2+1αpαδ,pα-1αpαδ-pα,pα=Ψ(α)+δ22α,

and the proof is complete. ∎

B Some convex analysis for index functions

We shall provide some additional details for convex index functions. First, it is well known that for a convex index function f we have that 0<st yields f(s)/sf(t)/t. Indeed, we let 0<θ:=st1 and obtain that

f(s)=f(θt+(1-θ)0)θf(t)+(1-θ)f(0)=stf(t),

which allows us to prove the assertion. This implies that the limit g:=limt0f(t)/t0 exists. If g>0, then f is linear near zero, and this case is not interesting in this study. Otherwise, we assume that g=0. In this interesting (sub-linear) case the following result is relevant.

Lemma B.1.

Suppose that f is a convex index function. The following assertions are equivalent.

  1. The quotient f(t)/t, t>0, is a strictly increasing index function.

  2. There is a strictly increasing index function φ , and the companion Θ(t):=tφ(t), t>0, such that the representation f(t)=Θ2((φ2)-1(t)), t>0, is valid.

Proof.

Clearly, if f has a representation as in (ii), then we find, letting φ2(s)=t, that

f(t)t=Θ2(s)φ2(s)=s0,

as s0.

For the other implication we observe that by assumption we can (implicitly) define the strictly increasing index function φ by

φ(f(t)t):=t,t>0.

This yields that

f(t)=t(φ2)-1(t)=Θ2((φ2)-1(t)),t>0,

which completes the proof. ∎

As an interesting consequence we mention the following result for the Fenchel conjugate function f to the convex (index) function f.

Corollary B.2.

Suppose that f is a convex index function such that the quotient f(t)/t, t>0, is a strictly increasing index function. Then the Fenchel conjugate function f is an index function, and there is a strictly increasing index function φ such that

f(t)tφ2(t),t>0.

Proof.

First, by Lemma B.1 there is a strictly increasing index function φ such that f(t)=Θ2((φ2)-1(t)), t>0. Now we use the “poor man’s Young Inequality” of the form

φ2(x)yφ2(x)x+φ2(y)y,x,y>0,

which is easily seen by considering the cases x<y and yx separately. This in turn, by letting s:=φ2(x) and t:=y, implies

stΘ2((φ2)-1(s))+Θ2(t),s,t>0.

For the Fenchel conjugate f this yields

f(t):=sups>0{st-f(s)}Θ2(t).

From this bound we conclude that f will be an index function for which the quotient f(t)/t has the desired bound. ∎

Acknowledgements

We thank Peter Elbau (University of Vienna) and Jens Flemming (TU Chemnitz) for suggesting to us essential ingredients for the proofs of Propositions 2.6 and 3.2, respectively.

References

[1] V. Albani, P. Elbau, M. V. de Hoop and O. Scherzer, Optimal convergence rates results for linear inverse problems in Hilbert spaces, Numer. Funct. Anal. Optim. 37 (2016), no. 5, 521–540. 10.1080/01630563.2016.1144070Search in Google Scholar PubMed PubMed Central

[2] R. Andreev, P. Elbau, M. V. de Hoop, L. Qiu and O. Scherzer, Generalized convergence rates results for linear inverse problems in Hilbert spaces, Numer. Funct. Anal. Optim. 36 (2015), no. 5, 549–566. 10.1080/01630563.2015.1021422Search in Google Scholar

[3] S. W. Anzengruber, B. Hofmann and P. Mathé, Regularization properties of the sequential discrepancy principle for Tikhonov regularization in Banach spaces, Appl. Anal. 93 (2014), no. 7, 1382–1400. 10.1080/00036811.2013.833326Search in Google Scholar

[4] M. Burger, J. Flemming and B. Hofmann, Convergence rates in 1-regularization if the sparsity assumption fails, Inverse Problems 29 (2013), no. 2, Article ID 025013. 10.1088/0266-5611/29/2/025013Search in Google Scholar

[5] M. Burger and S. Osher, Convergence rates of convex variational regularization, Inverse Problems 20 (2004), no. 5, 1411–1421. 10.1088/0266-5611/20/5/005Search in Google Scholar

[6] A. Chambolle, V. Caselles, D. Cremers, M. Novaga and T. Pock, An introduction to total variation for image analysis, Theoretical Foundations and Numerical Methods for Sparse Recovery, Radon Ser. Comput. Appl. Math. 9, Walter de Gruyter, Berlin (2010), 263–340. 10.1515/9783110226157.263Search in Google Scholar

[7] H. W. Engl, M. Hanke and A. Neubauer, Regularization of Inverse Problems, Math. Appl. 375, Kluwer Academic, Dordrecht, 1996. 10.1007/978-94-009-1740-8Search in Google Scholar

[8] J. Flemming, Generalized Tikhonov Regularization and Modern Convergence Rate Theory in Banach Spaces, Shaker, Aachen, 2012. Search in Google Scholar

[9] J. Flemming, A converse result for Banach space convergence rates in Tikhonov-type convex regularization of ill-posed linear equations, J. Inverse Ill-Posed Probl. 26 (2018), no. 5, 639–646. 10.1515/jiip-2017-0116Search in Google Scholar

[10] J. Flemming, Existence of variational source conditions for nonlinear inverse problems in Banach spaces, J. Inverse Ill-Posed Probl. 26 (2018), no. 2, 277–286. 10.1515/jiip-2017-0092Search in Google Scholar

[11] J. Flemming and D. Gerth, Injectivity and weak*-to-weak continuity suffice for convergence rates in 1-regularization, J. Inverse Ill-Posed Probl. 26 (2018), no. 1, 85–94. 10.1515/jiip-2017-0008Search in Google Scholar

[12] J. Flemming and B. Hofmann, A new approach to source conditions in regularization with general residual term, Numer. Funct. Anal. Optim. 31 (2010), no. 1–3, 254–284. 10.1080/01630561003765721Search in Google Scholar

[13] J. Flemming, B. Hofmann and P. Mathé, Sharp converse results for the regularization error using distance functions, Inverse Problems 27 (2011), no. 2, Article ID 025006. 10.1088/0266-5611/27/2/025006Search in Google Scholar

[14] K. Frick, D. A. Lorenz and E. Resmerita, Morozov’s principle for the augmented Lagrangian method applied to linear inverse problems, Multiscale Model. Simul. 9 (2011), no. 4, 1528–1548. 10.1137/100812835Search in Google Scholar

[15] M. Grasmair, Generalized Bregman distances and convergence rates for non-convex regularization methods, Inverse Problems 26 (2010), no. 11, Article ID 115014. 10.1088/0266-5611/26/11/115014Search in Google Scholar

[16] B. Hofmann, B. Kaltenbacher, C. Pöschl and O. Scherzer, A convergence rates result for Tikhonov regularization in Banach spaces with non-smooth operators, Inverse Problems 23 (2007), no. 3, 987–1010. 10.1088/0266-5611/23/3/009Search in Google Scholar

[17] B. Hofmann and P. Mathé, Analysis of profile functions for general linear regularization methods, SIAM J. Numer. Anal. 45 (2007), no. 3, 1122–1141. 10.1137/060654530Search in Google Scholar

[18] B. Hofmann and P. Mathé, Parameter choice in Banach space regularization under variational inequalities, Inverse Problems 28 (2012), Article ID 104006. 10.1088/0266-5611/28/10/104006Search in Google Scholar

[19] T. Hohage and F. Weidling, Characterizations of variational source conditions, converse results, and maxisets of spectral regularization methods, SIAM J. Numer. Anal. 55 (2017), no. 2, 598–620. 10.1137/16M1067445Search in Google Scholar

[20] K. Ito and B. Jin, Inverse Problems. Tikhonov Theory and Algorithms, Ser. Appl. Math. 22, World Scientific, Hackensack 2015. 10.1142/9120Search in Google Scholar

[21] S. Kindermann, Convex Tikhonov regularization in Banach spaces: New results on convergence rates, J. Inverse Ill-Posed Probl. 24 (2016), no. 3, 341–350. 10.1515/jiip-2015-0038Search in Google Scholar

[22] P. Mathé and S. V. Pereverzev, Geometry of linear ill-posed problems in variable Hilbert scales, Inverse Problems 19 (2003), no. 3, 789–803. 10.1088/0266-5611/19/3/319Search in Google Scholar

[23] Y. Meyer, Oscillating Patterns in Image Processing and Nonlinear Evolution Equations, Univ. Lecture Ser. 22, American Mathematical Society, Providence, 2001. 10.1090/ulect/022Search in Google Scholar

[24] A. Neubauer, T. Hein, B. Hofmann, S. Kindermann and U. Tautenhahn, Improved and extended results for enhanced convergence rates of Tikhonov regularization in Banach spaces, Appl. Anal. 89 (2010), no. 11, 1729–1743. 10.1080/00036810903517597Search in Google Scholar

[25] E. Resmerita, Regularization of ill-posed problems in Banach spaces: Convergence rates, Inverse Problems 21 (2005), no. 4, 1303–1314. 10.1088/0266-5611/21/4/007Search in Google Scholar

[26] E. Resmerita and O. Scherzer, Error estimates for non-quadratic regularization and the relation to enhancement, Inverse Problems 22 (2006), no. 3, 801–814. 10.1088/0266-5611/22/3/004Search in Google Scholar

[27] L. I. Rudin, S. Osher and E. Fatemi, Nonlinear total variation based noise removal algorithms, Phys. D 60 (1992), no. 1–4, 259–268. 10.1016/0167-2789(92)90242-FSearch in Google Scholar

[28] O. Scherzer, M. Grasmair, H. Grossauer, M. Haltmeier and F. Lenzen, Variational Methods in Imaging, Appl. Math. Sci. 167, Springer, New York, 2009. Search in Google Scholar

[29] T. Schuster, B. Kaltenbacher, B. Hofmann and K. S. Kazimierski, Regularization Methods in Banach Spaces, Radon Ser. Comput. Appl. Math. 10, Walter de Gruyter, Berlin, 2012. 10.1515/9783110255720Search in Google Scholar

Received: 2018-05-03
Revised: 2019-01-31
Accepted: 2019-02-08
Published Online: 2019-03-08
Published in Print: 2019-04-01

© 2019 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 10.10.2025 from https://www.degruyterbrill.com/document/doi/10.1515/jiip-2018-0039/html
Scroll to top button