Home Weighting operators for sparsity regularization
Article
Licensed
Unlicensed Requires Authentication

Weighting operators for sparsity regularization

  • Ole Løseth Elvetun , Bjørn Fredrik Nielsen and Niranjana Sudheer EMAIL logo
Published/Copyright: October 1, 2025

Abstract

Standard regularization methods typically favor solutions which are in, or close to, the orthogonal complement of the null space of the forward operator/matrix 𝖠 . This particular biasedness might not be desirable in applications and can lead to severe challenges when 𝖠 is non-injective. We have therefore, in a series of papers, investigated how to “remedy” this fact, relative to a chosen basis and in a certain mathematical sense: Based on a weighting procedure, it turns out that it is possible to modify both Tikhonov and sparsity regularization such that each member of the chosen basis can be almost perfectly recovered from their image under 𝖠 . In particular, we have studied this problem for the task of using boundary data to identify the source term in an elliptic PDE. However, this weighting procedure involves 𝖠 𝖠 , where 𝖠 denotes the pseudo-inverse of 𝖠 , and can thus be CPU-demanding and lead to undesirable error amplification. We therefore, in this paper, study alternative weighting approaches and prove that some of the recovery results established for the methodology involving 𝖠 hold for a broader class of weighting schemes. In fact, it turns out that “any” linear operator 𝖡 has an associated proper weighting defined in terms of images under 𝖡𝖠 . We also present a series of numerical experiments, employing different choices of 𝖡 .

MSC 2020: 65N21; 65F22; 65K10

A Proof of Theorem 3.3

Using the notation

h ( 𝐳 ) = 𝐳 1 ,

the first order optimality condition for (3.7) reads

𝟎 𝖢 T ( 𝖢 𝐱 - 𝖢 𝐞 j ) + α 𝖶 h ( 𝖶 𝐱 ) ,

where h denotes the subgradient of h. The involved cost-functional is convex, and this condition is thus both necessary and sufficient. Inserting 𝐱 = γ α 𝐞 j to the condition above, we obtain

( 1 - γ α ) 𝖢 T 𝖢 𝐞 j α 𝖶 h ( γ α 𝖶 𝐞 j ) ,

or, alternatively,

(A.1) ( 1 - γ α ) α 𝖶 - 1 𝖢 T 𝖢 𝐞 j h ( γ α 𝖶 𝐞 j ) .

Since the entries of the diagonal matrix 𝖶 are strictly positive, it follows from standard computations that

( h ( γ α 𝖶 𝐞 j ) , 𝐞 i ) = { 1 , i = j , , i j ,

provided that γ α > 0 . We thus may write (A.1) in the following form:

(A.2) ( 1 - γ α ) α ( 𝖶 - 1 𝖢 T 𝖢 𝐞 j , 𝐞 i ) { 1 , i = j , , i j .

Invoking (3.6) we therefore obtain the requirement

(A.3) ( 1 - γ α ) α 𝖢 𝐞 j ( 𝖢 𝐞 j 𝖢 𝐞 j , 𝖢 𝐞 i 𝖢 𝐞 i ) { 1 , i = j , , i j .

With the choice

γ α = 1 - α w j ,

it follows that (A.3) holds for i = j , by recalling that w j = 𝖢 𝐞 j . From the Cauchy–Schwarz inequality, we observe that (A.3) also is satisfied when i j , proving existence of the minimizer.

To show uniqueness, we first denote the cost-functional by 𝔍 , i.e.,

𝔍 ( 𝐱 ) = 1 2 𝖢 𝐱 - 𝖢 𝐞 j 2 + α 𝖶 𝐱 1 .

Let 𝐲 n , 𝐲 𝐱 α be arbitrary. We will show that no such 𝐲 can be a minimizer, i.e., the minimizer is unique. We split the analysis into two cases:

Case 1: 𝐲 = c 𝐱 α , c 1 .

By the convexity of the cost-functional in (3.7) and the argument presented above, it follows that 𝐲 = c 𝐱 α cannot be a minimizer unless c = 1 .

Case 2: 𝐲 c 𝐱 α .

In this case there must exist at least one component y k , k j , of 𝐲 such that y k 0 . Consider

𝔍 ( 𝐲 ) - 𝔍 ( 𝐱 α ) = 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + α ( 𝖶 𝐲 1 - 𝖶 𝐱 α 1 ) .

Also, by the definition of the subdifferential,

h ( 𝖶 𝐲 ) - h ( 𝖶 𝐱 α ) 𝐳 T ( 𝖶 𝐲 - 𝖶 𝐱 α )

for any 𝐳 h ( 𝖶 𝐱 α ) . Consequently, we get

𝔍 ( 𝐲 ) - 𝔍 ( 𝐱 α ) = 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + α ( h ( 𝖶 𝐲 ) - h ( 𝖶 𝐱 α ) )
(A.4) 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + α 𝐳 T ( 𝖶 𝐲 - 𝖶 𝐱 α ) .

Recall that 𝐱 α = γ α 𝐞 j . From Lemma (3.2) , we can write (A.2) as

(A.5)

1 α ( 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) , 𝐞 i ) { 1 , i = j , ( - 1 , 1 ) , i j ,
{ 1 , i = j , , i j ,
= ( h ( 𝖶 𝐱 α ) , 𝐞 i ) .

This implies that

(A.6) 1 α 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) h ( 𝖶 𝐱 α ) .

However, choosing 𝐳 = 1 α 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) does not immediately lead to a strict inequality in (A.4). Consequently, we must find a better choice of 𝐳 . Without loss of generality,[1] we can assume that [ 𝖶 𝐲 - 𝖶 𝐱 α ] k > 0 and choose 𝐳 ~ = [ z ~ 1 , z ~ 2 , , z ~ n ] T , where z ~ i is defined as

z ~ i = { 1 , i = k , 1 α ( 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) , 𝐞 i ) , i k .

Since the condition (A.6) holds, it follows that 𝐳 ~ h ( 𝖶 𝐱 α ) .

From (A.5) we have [ 1 α 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) ] k < 1 and therefore we get the strictly inequality

𝐳 ~ T ( 𝖶 𝐲 - 𝖶 𝐱 α ) > 1 α 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) T ( 𝖶 𝐲 - 𝖶 𝐱 α ) .

Finally, combining this inequality with (A.4), we obtain

𝔍 ( 𝐲 ) - 𝔍 ( 𝐱 α ) 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + α 𝐳 ~ T ( 𝖶 𝐲 - 𝖶 𝐱 α )
> 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + ( 𝖶 - 1 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) ) T ( 𝖶 𝐲 - 𝖶 𝐱 α )
= 1 2 𝖢 𝐲 - 𝖢 𝐞 j 2 - 1 2 𝖢 𝐱 α - 𝖢 𝐞 j 2 + ( 𝖢 T 𝖢 ( 𝐞 j - 𝐱 α ) ) T ( 𝐲 - 𝐱 α )
0 ,

where the final inequality follows from the first-order optimality conditions of the convex functional

g ( 𝐱 ) = 1 2 𝖢 𝐱 - 𝖢 𝐞 j 2 ,

i.e.,

g ( 𝐲 ) - g ( 𝐱 α ) g ( 𝐱 α ) T ( 𝐲 - 𝐱 α ) = ( 𝖢 T 𝖢 ( 𝐱 α - 𝐞 j ) ) T ( 𝐲 - 𝐱 α ) .

This shows that 𝐱 α is the unique minimizer of 𝔍 ( 𝐱 ) .

B Proof of Theorem 3.5

Let

𝐜 = j 𝒥 sgn ( x j * ) 𝖢 𝐞 j 𝖢 𝐞 j .

If we can show that (3.8) and (3.9) hold for this choice of 𝐜 , Theorem 3.5 will follow immediately from Theorem 3.4.

For i 𝒥 , we have from the orthogonality (3.13) of { 𝖢 𝐞 j } j 𝒥 that

𝖢 𝐞 i 𝖢 𝐞 i 𝐜 = 𝖢 𝐞 i 𝖢 𝐞 i 𝖢 𝐞 i 𝖢 𝐞 i sgn ( x i * ) = sgn ( x i * ) ,

which shows that (3.8) holds.

For i 𝒥 c , the support assumption (3.11) implies that we have at most one k 𝒥 such that i supp ( 𝖢 T 𝖢 𝐞 k ) . Consequently,

𝖢 𝐞 i 𝖢 𝐞 i 𝐜 = j 𝒥 sgn ( x j * ) 𝖢 𝐞 i 𝖢 𝐞 j 𝖢 𝐞 i 𝖢 𝐞 j
= j 𝒥 sgn ( x j * ) 𝐞 i 𝖢 T 𝖢 𝐞 j 𝖢 𝐞 i 𝖢 𝐞 j
= sgn ( x k * ) 𝐞 i 𝖢 T 𝖢 𝐞 k 𝖢 𝐞 i 𝖢 𝐞 k
(B.1) = sgn ( x k * ) 𝖢 𝐞 i 𝖢 𝐞 k 𝖢 𝐞 i 𝖢 𝐞 k .

Invoking the Cauchy–Schwarz inequality, it follows that

| 𝖢 𝐞 i 𝖢 𝐞 k | < 𝖢 𝐞 i 𝖢 𝐞 k ,

where the strict inequality can be asserted from the non-parallelism assumption (2.5). Inserting this in (B.1) gives

| 𝖢 𝐞 i 𝖢 𝐞 i 𝐜 | < 1 ,

which shows that also condition (3.9) of Theorem 3.4 is satisfied.

On the other hand, if i 𝒥 c and i supp ( 𝖢 T 𝖢 𝐞 j ) for any j 𝒥 , we get that

𝖢 𝐞 i 𝖢 𝐞 i 𝐜 = 0 ,

showing that the condition (3.9) also holds in this case. Thus, we can conclude that 𝐱 * is a solution to the problem (3.12).

To prove the uniqueness, assume that there exists another minimizer 𝐲 . Since both (3.8) and (3.9) are shown to hold, it follows from Theorem 3.4 that supp ( 𝐲 ) supp ( 𝐱 * ) . Consequently, we can write

𝖠 𝐱 α = 𝖠 𝐲

in the form

𝖠 j 𝒥 y j 𝐞 j = 𝖠 j 𝒥 x j * 𝐞 j .

Furthermore, we can multiply with 𝖡 to obtain

j 𝒥 y j 𝖢 𝐞 j = j 𝒥 x j * 𝖢 𝐞 j .

The orthogonality of { 𝖢 𝐞 j } j 𝒥 ensures that y j must be equal to x j * for all j 𝒥 , which implies uniqueness.

References

[1] D. Calvetti, Preconditioned iterative methods for linear discrete ill-posed problems from a Bayesian inversion perspective, J. Comput. Appl. Math. 198 (2007), no. 2, 378–395. 10.1016/j.cam.2005.10.038Search in Google Scholar

[2] D. Calvetti and E. Somersalo, Inverse problems: From regularization to Bayesian inference, Wiley Interdiscip. Rev. Comput. Stat. 10 (2018), no. 3, Article ID e1427. 10.1002/wics.1427Search in Google Scholar

[3] E. J. Candes and T. Tao, Decoding by linear programming, IEEE Trans. Inform. Theory 51 (2005), no. 12, 4203–4215. 10.1109/TIT.2005.858979Search in Google Scholar

[4] E. J. Candès, M. B. Wakin and S. P. Boyd, Enhancing sparsity by reweighted l 1 minimization, J. Fourier Anal. Appl. 14 (2008), no. 5–6, 877–905. 10.1007/s00041-008-9045-xSearch in Google Scholar

[5] R. G. de Peralta Menendez, O. Hauk, S. Gonzalez Andino, H. Vogt and C. Michel, Linear inverse solutions with optimal resolution kernels applied to electromagnetic tomography, Human Brain Mapping 5 (1997), no. 6, 454–467. 10.1002/(SICI)1097-0193(1997)5:6<454::AID-HBM6>3.0.CO;2-2Search in Google Scholar

[6] D. L. Donoho and M. Elad, Optimally sparse representation in general (nonorthogonal) dictionaries via l 1 minimization, Proc. Natl. Acad. Sci. USA 100 (2003), no. 5, 2197–2202. 10.1073/pnas.0437847100Search in Google Scholar

[7] V. Duval and G. Peyré, Sparse regularization on thin grids I: the Lasso, Inverse Problems 33 (2017), no. 5, Article ID 055008. 10.1088/1361-6420/aa5e12Search in Google Scholar

[8] O. L. Elvetun and B. F. Nielsen, A regularization operator for source identification for elliptic PDEs, Inverse Probl. Imaging 15 (2021), no. 4, 599–618. 10.3934/ipi.2021006Search in Google Scholar

[9] O. L. Elvetun and B. F. Nielsen, Weighted sparsity regularization for source identification for elliptic PDEs, J. Inverse Ill-Posed Probl. 31 (2023), no. 5, 687–709. 10.1515/jiip-2021-0057Search in Google Scholar

[10] O. L. Elvetun and B. F. Nielsen, Box constraints and weighted sparsity regularization for identifying sources in elliptic PDEs, Numer. Funct. Anal. Optim. 45 (2024), no. 16, 779–812. 10.1080/01630563.2024.2405489Search in Google Scholar

[11] O. L. Elvetun and B. F. Nielsen, Identifying the source term in the potential equation with weighted sparsity regularization, Math. Comp. 93 (2024), no. 350, 2811–2836. 10.1090/mcom/3941Search in Google Scholar

[12] J.-J. Fuchs, On sparse representations in arbitrary redundant bases, IEEE Trans. Inform. Theory 50 (2004), no. 6, 1341–1344. 10.1109/TIT.2004.828141Search in Google Scholar

[13] M. Fuchs, M. Wagner, T. Köhler and H.-A. Wischmann, Linear and nonlinear current density reconstructions, J. Clin. Neurophys. 16 (1999), no. 3, 267–295. 10.1097/00004691-199905000-00006Search in Google Scholar PubMed

[14] H. Garde and K. Knudsen, 3D reconstruction for partial data electrical impedance tomography using a sparsity prior, Discrete Contin. Dyn. Syst. 2015 (2015), 495–504. 10.3934/proc.2015.0495Search in Google Scholar

[15] I. F. Gorodnitsky, J. S. George and B. D. Rao, Neuromagnetic source imaging with FOCUSS: A recursive weighted minimum norm algorithm, Electroencephalography Clinical Neurophy. 95 (1995), no. 4, 231–251. 10.1016/0013-4694(95)00107-ASearch in Google Scholar PubMed

[16] M. Grasmair, M. Haltmeier and O. Scherzer, Necessary and sufficient conditions for linear convergence of 1 -regularization, Comm. Pure Appl. Math. 64 (2011), no. 2, 161–182. 10.1002/cpa.20350Search in Google Scholar

[17] F.-H. Lin, T. Witzel, S. P. Ahlfors, S. M. Stufflebeam, J. W. Belliveau and M. S. Hämäläinen, Assessing and improving the spatial accuracy in MEG source localization by depth-weighted minimum-norm estimates, NeuroImage 31 (2006), no. 1, 160–171. 10.1016/j.neuroimage.2005.11.054Search in Google Scholar PubMed

[18] A. Logg, K.-A. Mardal and G. N. Wells, Automated Solution of Differential Equations by the Finite Element Method, Lect. Notes Comput. Sci. Eng. 84, Springer, Heidelberg, 2012. 10.1007/978-3-642-23099-8Search in Google Scholar

[19] F. Lucka, S. Pursiainen, M. Burger and C. H. Wolters, Hierarchical Bayesian inference for the EEG inverse problem using realistic FE head models: Depth localization and source separation for focal primary currents, NeuroImage 61 (2012), no. 4, 1364–1382. 10.1016/j.neuroimage.2012.04.017Search in Google Scholar PubMed

[20] T. Lyche, Numerical Linear Algebra and Matrix Factorizations, Texts Comput. Sci. Eng. 22, Springer, Cham, 2020. 10.1007/978-3-030-36468-7Search in Google Scholar

[21] R. D. Pasqual-Marqui, Review of methods for solving the EEG inverse problem, Int. J. Bioelectromagnetism 1 (1999), no. 1, 75–86. Search in Google Scholar

[22] R. D. Pasqual-Marqui, Standardized low-resolution brain electromagnetic tomography (sLORETA): Technical details, Methods Find. Exp. Clin. Pharmacol. 24 (2002), 5–12. Search in Google Scholar

[23] R. Tibshirani, Regression shrinkage and selection via the lasso, J. Roy. Statist. Soc. Ser. B 58 (1996), no. 1, 267–288. 10.1111/j.2517-6161.1996.tb02080.xSearch in Google Scholar

[24] J. A. Tropp, Greed is good: Algorithmic results for sparse approximation, IEEE Trans. Inform. Theory 50 (2004), no. 10, 2231–2242. 10.1109/TIT.2004.834793Search in Google Scholar

[25] P. Xu, Y. Tian, H. Chen and D. Yao, Lp norm iterative sparse solution for EEG source localization, IEEE Trans. Biomed. Eng. 54 (2007), no. 3, 400–409. 10.1109/TBME.2006.886640Search in Google Scholar PubMed

Received: 2025-05-08
Revised: 2025-09-02
Accepted: 2025-09-04
Published Online: 2025-10-01

© 2025 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 23.10.2025 from https://www.degruyterbrill.com/document/doi/10.1515/jiip-2025-0033/html
Scroll to top button