Startseite Mathematik Weighting operators for sparsity regularization
Artikel
Lizenziert
Nicht lizenziert Erfordert eine Authentifizierung

Weighting operators for sparsity regularization

  • Ole Løseth Elvetun , Bjørn Fredrik Nielsen und Niranjana Sudheer EMAIL logo
Veröffentlicht/Copyright: 1. Oktober 2025

Abstract

Standard regularization methods typically favor solutions which are in, or close to, the orthogonal complement of the null space of the forward operator/matrix 𝖠 . This particular biasedness might not be desirable in applications and can lead to severe challenges when 𝖠 is non-injective. We have therefore, in a series of papers, investigated how to “remedy” this fact, relative to a chosen basis and in a certain mathematical sense: Based on a weighting procedure, it turns out that it is possible to modify both Tikhonov and sparsity regularization such that each member of the chosen basis can be almost perfectly recovered from their image under 𝖠 . In particular, we have studied this problem for the task of using boundary data to identify the source term in an elliptic PDE. However, this weighting procedure involves 𝖠 𝖠 , where 𝖠 denotes the pseudo-inverse of 𝖠 , and can thus be CPU-demanding and lead to undesirable error amplification. We therefore, in this paper, study alternative weighting approaches and prove that some of the recovery results established for the methodology involving 𝖠 hold for a broader class of weighting schemes. In fact, it turns out that “any” linear operator 𝖡 has an associated proper weighting defined in terms of images under 𝖡𝖠 . We also present a series of numerical experiments, employing different choices of 𝖡 .

MSC 2020: 65N21; 65F22; 65K10

A Proof of Theorem 3.3

Using the notation

h ( 𝐳 ) = 𝐳 1 ,

the first order optimality condition for (3.7) reads

𝟎 𝖢 T ( 𝖢 𝐱 - 𝖢 𝐞 j ) + α 𝖶 h ( 𝖶 𝐱 ) ,

where h denotes the subgradient of h. The involved cost-functional is convex, and this condition is thus both necessary and sufficient. Inserting 𝐱 = γ α 𝐞 j to the condition above, we obtain

( 1 - γ α ) 𝖢 T 𝖢 𝐞 j α 𝖶 h ( γ α 𝖶 𝐞 j ) ,

or, alternatively,

(A.1) ( 1 - γ α ) α 𝖶 - 1 𝖢 T 𝖢 𝐞 j h ( γ α 𝖶 𝐞 j ) .

Since the entries of the diagonal matrix 𝖶 are strictly positive, it follows from standard computations that

( h ( γ α 𝖶 𝐞 j ) , 𝐞 i ) = { 1 , i = j , , i j ,

provided that γ α > 0 . We thus may write (A.1) in the following form:

(A.2) ( 1 - γ α ) α ( 𝖶 - 1 𝖢 T 𝖢 𝐞 j , 𝐞 i ) { 1 , i = j , , i j .

Invoking (3.6) we therefore obtain the requirement

(A.3) ( 1 - γ α ) α 𝖢 𝐞 j ( 𝖢 𝐞 j 𝖢 𝐞 j , 𝖢 𝐞 i 𝖢 𝐞 i ) { 1 , i = j , , i j .

With the choice

γ α = 1 - α w j ,

it follows that (A.3) holds for i = j , by recalling that w j = 𝖢 𝐞 j . From the Cauchy–Schwarz inequality, we observe that (A.3) also is satisfied when i j , proving existence of the minimizer.

To show uniqueness, we first denote the cost-functional by 𝔍 , i.e.,

𝔍 ( 𝐱 ) = 1 2 𝖢 𝐱 - 𝖢 𝐞 j 2 + α 𝖶 𝐱 1 .

Let 𝐲 n , 𝐲 𝐱 α be arbitrary. We will show that no such 𝐲 can be a minimizer, i.e., the minimizer is unique. We split the analysis into two cases:

Case 1: 𝐲 = c 𝐱 α , c 1 .

By the convexity of the cost-functional in (3.7) and the argument presented above, it follows that 𝐲 = c 𝐱 α cannot be a minimizer unless c = 1 .

Case 2: 𝐲 c 𝐱 α .

In this case there must exist at least one component y k , k j , of 𝐲 such that y k 0 . Consider

𝔍 ( 𝐲 ) - 𝔍 ( 𝐱 α ) = 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + α ( 𝖶 𝐲 1 - 𝖶 𝐱 α 1 ) .

Also, by the definition of the subdifferential,

h ( 𝖶 𝐲 ) - h ( 𝖶 𝐱 α ) 𝐳 T ( 𝖶 𝐲 - 𝖶 𝐱 α )

for any 𝐳 h ( 𝖶 𝐱 α ) . Consequently, we get

𝔍 ( 𝐲 ) - 𝔍 ( 𝐱 α ) = 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + α ( h ( 𝖶 𝐲 ) - h ( 𝖶 𝐱 α ) )
(A.4) 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + α 𝐳 T ( 𝖶 𝐲 - 𝖶 𝐱 α ) .

Recall that 𝐱 α = γ α 𝐞 j . From Lemma (3.2) , we can write (A.2) as

(A.5)

1 α ( 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) , 𝐞 i ) { 1 , i = j , ( - 1 , 1 ) , i j ,
{ 1 , i = j , , i j ,
= ( h ( 𝖶 𝐱 α ) , 𝐞 i ) .

This implies that

(A.6) 1 α 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) h ( 𝖶 𝐱 α ) .

However, choosing 𝐳 = 1 α 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) does not immediately lead to a strict inequality in (A.4). Consequently, we must find a better choice of 𝐳 . Without loss of generality,[1] we can assume that [ 𝖶 𝐲 - 𝖶 𝐱 α ] k > 0 and choose 𝐳 ~ = [ z ~ 1 , z ~ 2 , , z ~ n ] T , where z ~ i is defined as

z ~ i = { 1 , i = k , 1 α ( 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) , 𝐞 i ) , i k .

Since the condition (A.6) holds, it follows that 𝐳 ~ h ( 𝖶 𝐱 α ) .

From (A.5) we have [ 1 α 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) ] k < 1 and therefore we get the strictly inequality

𝐳 ~ T ( 𝖶 𝐲 - 𝖶 𝐱 α ) > 1 α 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) T ( 𝖶 𝐲 - 𝖶 𝐱 α ) .

Finally, combining this inequality with (A.4), we obtain

𝔍 ( 𝐲 ) - 𝔍 ( 𝐱 α ) 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + α 𝐳 ~ T ( 𝖶 𝐲 - 𝖶 𝐱 α )
> 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + ( 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) ) T ( 𝖶 𝐲 - 𝖶 𝐱 α )
= 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + ( 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) ) T ( 𝐲 - 𝐱 α )
0 ,

where the final inequality follows from the first-order optimality conditions of the convex functional

g ( 𝐱 ) = 1 2 𝖢 𝐱 - 𝖢 𝐞 j 2 ,

i.e.,

g ( 𝐲 ) - g ( 𝐱 α ) g ( 𝐱 α ) T ( 𝐲 - 𝐱 α ) = ( 𝖢 T 𝖢 ( 𝐱 α - 𝐞 j ) ) T ( 𝐲 - 𝐱 α ) .

This shows that 𝐱 α is the unique minimizer of 𝔍 ( 𝐱 ) .

B Proof of Theorem 3.5

Let

𝐜 = j 𝒥 sgn ( x j * ) 𝖢 𝐞 j 𝖢 𝐞 j .

If we can show that (3.8) and (3.9) hold for this choice of 𝐜 , Theorem 3.5 will follow immediately from Theorem 3.4.

For i 𝒥 , we have from the orthogonality (3.13) of { 𝖢 𝐞 j } j 𝒥 that

𝖢 𝐞 i 𝖢 𝐞 i 𝐜 = 𝖢 𝐞 i 𝖢 𝐞 i 𝖢 𝐞 i 𝖢 𝐞 i sgn ( x i * ) = sgn ( x i * ) ,

which shows that (3.8) holds.

For i 𝒥 c , the support assumption (3.11) implies that we have at most one k 𝒥 such that i supp ( 𝖢 T 𝖢 𝐞 k ) . Consequently,

𝖢 𝐞 i 𝖢 𝐞 i 𝐜 = j 𝒥 sgn ( x j * ) 𝖢 𝐞 i 𝖢 𝐞 j 𝖢 𝐞 i 𝖢 𝐞 j
= j 𝒥 sgn ( x j * ) 𝐞 i 𝖢 T 𝖢 𝐞 j 𝖢 𝐞 i 𝖢 𝐞 j
= sgn ( x k * ) 𝐞 i 𝖢 T 𝖢 𝐞 k 𝖢 𝐞 i 𝖢 𝐞 k
(B.1) = sgn ( x k * ) 𝖢 𝐞 i 𝖢 𝐞 k 𝖢 𝐞 i 𝖢 𝐞 k .

Invoking the Cauchy–Schwarz inequality, it follows that

| 𝖢 𝐞 i 𝖢 𝐞 k | < 𝖢 𝐞 i 𝖢 𝐞 k ,

where the strict inequality can be asserted from the non-parallelism assumption (2.5). Inserting this in (B.1) gives

| 𝖢 𝐞 i 𝖢 𝐞 i 𝐜 | < 1 ,

which shows that also condition (3.9) of Theorem 3.4 is satisfied.

On the other hand, if i 𝒥 c and i supp ( 𝖢 T 𝖢 𝐞 j ) for any j 𝒥 , we get that

𝖢 𝐞 i 𝖢 𝐞 i 𝐜 = 0 ,

showing that the condition (3.9) also holds in this case. Thus, we can conclude that 𝐱 * is a solution to the problem (3.12).

To prove the uniqueness, assume that there exists another minimizer 𝐲 . Since both (3.8) and (3.9) are shown to hold, it follows from Theorem 3.4 that supp ( 𝐲 ) supp ( 𝐱 * ) . Consequently, we can write

𝖠 𝐱 α = 𝖠 𝐲

in the form

𝖠 j 𝒥 y j 𝐞 j = 𝖠 j 𝒥 x j * 𝐞 j .

Furthermore, we can multiply with 𝖡 to obtain

j 𝒥 y j 𝖢 𝐞 j = j 𝒥 x j * 𝖢 𝐞 j .

The orthogonality of { 𝖢 𝐞 j } j 𝒥 ensures that y j must be equal to x j * for all j 𝒥 , which implies uniqueness.

References

[1] D. Calvetti, Preconditioned iterative methods for linear discrete ill-posed problems from a Bayesian inversion perspective, J. Comput. Appl. Math. 198 (2007), no. 2, 378–395. 10.1016/j.cam.2005.10.038Suche in Google Scholar

[2] D. Calvetti and E. Somersalo, Inverse problems: From regularization to Bayesian inference, Wiley Interdiscip. Rev. Comput. Stat. 10 (2018), no. 3, Article ID e1427. 10.1002/wics.1427Suche in Google Scholar

[3] E. J. Candes and T. Tao, Decoding by linear programming, IEEE Trans. Inform. Theory 51 (2005), no. 12, 4203–4215. 10.1109/TIT.2005.858979Suche in Google Scholar

[4] E. J. Candès, M. B. Wakin and S. P. Boyd, Enhancing sparsity by reweighted l 1 minimization, J. Fourier Anal. Appl. 14 (2008), no. 5–6, 877–905. 10.1007/s00041-008-9045-xSuche in Google Scholar

[5] R. G. de Peralta Menendez, O. Hauk, S. Gonzalez Andino, H. Vogt and C. Michel, Linear inverse solutions with optimal resolution kernels applied to electromagnetic tomography, Human Brain Mapping 5 (1997), no. 6, 454–467. 10.1002/(SICI)1097-0193(1997)5:6<454::AID-HBM6>3.0.CO;2-2Suche in Google Scholar

[6] D. L. Donoho and M. Elad, Optimally sparse representation in general (nonorthogonal) dictionaries via l 1 minimization, Proc. Natl. Acad. Sci. USA 100 (2003), no. 5, 2197–2202. 10.1073/pnas.0437847100Suche in Google Scholar

[7] V. Duval and G. Peyré, Sparse regularization on thin grids I: the Lasso, Inverse Problems 33 (2017), no. 5, Article ID 055008. 10.1088/1361-6420/aa5e12Suche in Google Scholar

[8] O. L. Elvetun and B. F. Nielsen, A regularization operator for source identification for elliptic PDEs, Inverse Probl. Imaging 15 (2021), no. 4, 599–618. 10.3934/ipi.2021006Suche in Google Scholar

[9] O. L. Elvetun and B. F. Nielsen, Weighted sparsity regularization for source identification for elliptic PDEs, J. Inverse Ill-Posed Probl. 31 (2023), no. 5, 687–709. 10.1515/jiip-2021-0057Suche in Google Scholar

[10] O. L. Elvetun and B. F. Nielsen, Box constraints and weighted sparsity regularization for identifying sources in elliptic PDEs, Numer. Funct. Anal. Optim. 45 (2024), no. 16, 779–812. 10.1080/01630563.2024.2405489Suche in Google Scholar

[11] O. L. Elvetun and B. F. Nielsen, Identifying the source term in the potential equation with weighted sparsity regularization, Math. Comp. 93 (2024), no. 350, 2811–2836. 10.1090/mcom/3941Suche in Google Scholar

[12] J.-J. Fuchs, On sparse representations in arbitrary redundant bases, IEEE Trans. Inform. Theory 50 (2004), no. 6, 1341–1344. 10.1109/TIT.2004.828141Suche in Google Scholar

[13] M. Fuchs, M. Wagner, T. Köhler and H.-A. Wischmann, Linear and nonlinear current density reconstructions, J. Clin. Neurophys. 16 (1999), no. 3, 267–295. 10.1097/00004691-199905000-00006Suche in Google Scholar PubMed

[14] H. Garde and K. Knudsen, 3D reconstruction for partial data electrical impedance tomography using a sparsity prior, Discrete Contin. Dyn. Syst. 2015 (2015), 495–504. 10.3934/proc.2015.0495Suche in Google Scholar

[15] I. F. Gorodnitsky, J. S. George and B. D. Rao, Neuromagnetic source imaging with FOCUSS: A recursive weighted minimum norm algorithm, Electroencephalography Clinical Neurophy. 95 (1995), no. 4, 231–251. 10.1016/0013-4694(95)00107-ASuche in Google Scholar PubMed

[16] M. Grasmair, M. Haltmeier and O. Scherzer, Necessary and sufficient conditions for linear convergence of 1 -regularization, Comm. Pure Appl. Math. 64 (2011), no. 2, 161–182. 10.1002/cpa.20350Suche in Google Scholar

[17] F.-H. Lin, T. Witzel, S. P. Ahlfors, S. M. Stufflebeam, J. W. Belliveau and M. S. Hämäläinen, Assessing and improving the spatial accuracy in MEG source localization by depth-weighted minimum-norm estimates, NeuroImage 31 (2006), no. 1, 160–171. 10.1016/j.neuroimage.2005.11.054Suche in Google Scholar PubMed

[18] A. Logg, K.-A. Mardal and G. N. Wells, Automated Solution of Differential Equations by the Finite Element Method, Lect. Notes Comput. Sci. Eng. 84, Springer, Heidelberg, 2012. 10.1007/978-3-642-23099-8Suche in Google Scholar

[19] F. Lucka, S. Pursiainen, M. Burger and C. H. Wolters, Hierarchical Bayesian inference for the EEG inverse problem using realistic FE head models: Depth localization and source separation for focal primary currents, NeuroImage 61 (2012), no. 4, 1364–1382. 10.1016/j.neuroimage.2012.04.017Suche in Google Scholar PubMed

[20] T. Lyche, Numerical Linear Algebra and Matrix Factorizations, Texts Comput. Sci. Eng. 22, Springer, Cham, 2020. 10.1007/978-3-030-36468-7Suche in Google Scholar

[21] R. D. Pasqual-Marqui, Review of methods for solving the EEG inverse problem, Int. J. Bioelectromagnetism 1 (1999), no. 1, 75–86. Suche in Google Scholar

[22] R. D. Pasqual-Marqui, Standardized low-resolution brain electromagnetic tomography (sLORETA): Technical details, Methods Find. Exp. Clin. Pharmacol. 24 (2002), 5–12. Suche in Google Scholar

[23] R. Tibshirani, Regression shrinkage and selection via the lasso, J. Roy. Statist. Soc. Ser. B 58 (1996), no. 1, 267–288. 10.1111/j.2517-6161.1996.tb02080.xSuche in Google Scholar

[24] J. A. Tropp, Greed is good: Algorithmic results for sparse approximation, IEEE Trans. Inform. Theory 50 (2004), no. 10, 2231–2242. 10.1109/TIT.2004.834793Suche in Google Scholar

[25] P. Xu, Y. Tian, H. Chen and D. Yao, Lp norm iterative sparse solution for EEG source localization, IEEE Trans. Biomed. Eng. 54 (2007), no. 3, 400–409. 10.1109/TBME.2006.886640Suche in Google Scholar PubMed

Received: 2025-05-08
Revised: 2025-09-02
Accepted: 2025-09-04
Published Online: 2025-10-01

© 2025 Walter de Gruyter GmbH, Berlin/Boston

Heruntergeladen am 8.12.2025 von https://www.degruyterbrill.com/document/doi/10.1515/jiip-2025-0033/html
Button zum nach oben scrollen