Home Local convergence of the error-reduction algorithm for real-valued objects
Article
Licensed
Unlicensed Requires Authentication

Local convergence of the error-reduction algorithm for real-valued objects

  • Gi-Ren Liu ORCID logo EMAIL logo
Published/Copyright: November 15, 2024

Abstract

Fourier phasing is the problem of retrieving Fourier phase information from Fourier intensity data. The error-reduction (ER) algorithm consists of two projections on the subspaces generated by the Fourier magnitude constraint and the object-domain constraint. The random phase illumination (RPI) and the real-valued constraint on the object significantly reduce the complexity of the intersection of the two subspaces. In this paper, we study how to approximate the projection of the starting point onto the subspace generated by the Fourier magnitude constraint by its projection on the tangent plane and estimate the approximation error by orthogonal decompositions. Moreover, we prove that the local geometric convergence rate of the ER algorithm is less than one almost surely and can be characterized as the cosine of the angle between the two projection spaces. A theoretical estimate of the convergence rate is derived and validated by some numerical experiments.

Funding statement: This work was supported by the National Center for Theoretical Sciences (NCTS, Taiwan) and the National Science and Technology Council under grant numbers 110-2628-M-006-003-MY3 and 113-2628-M-006-002-MY3.

A Derivation of (3.22)

We first observe that

r ( θ ) := 𝐠 - 𝐲 - i 𝐤 𝒦 ρ θ 𝐤 μ 𝐟 𝐤 𝐟 𝐤 * μ * 𝐲 2
= [ 𝐠 - 𝐲 - i 𝐤 𝒦 ρ θ 𝐤 μ 𝐟 𝐤 𝐟 𝐤 * μ * 𝐲 ] * [ 𝐠 - 𝐲 - i 𝐤 𝒦 ρ θ 𝐤 μ 𝐟 𝐤 𝐟 𝐤 * μ * 𝐲 ]
= 𝐠 - 𝐲 2 + 𝐤 𝒦 ρ θ 𝐤 μ 𝐟 𝐤 𝐟 𝐤 * μ * 𝐲 2 + 2 Re [ ( - i 𝐤 𝒦 ρ θ 𝐤 μ 𝐟 𝐤 𝐟 𝐤 * μ * 𝐲 ) * ( 𝐠 - 𝐲 ) ]
= ( 3.20 ) 𝐠 - 𝐲 2 + 𝐤 𝒦 ρ θ 𝐤 2 b 𝐤 2 - 2 𝐤 𝒦 ρ θ 𝐤 Im [ 𝐲 * μ 𝐟 𝐤 𝐟 𝐤 * μ * ( 𝐠 - 𝐲 ) ]
= 𝐠 - 𝐲 2 + 𝐤 𝒦 ρ θ 𝐤 2 b 𝐤 2 - 2 𝐤 𝒦 ρ θ 𝐤 Im ( 𝐲 * μ 𝐟 𝐤 𝐟 𝐤 * μ * 𝐠 ) .

The minimum point { θ 𝐤 } k 𝒦 ρ of the function r satisfies

0 = r θ 𝐤 = 2 θ 𝐤 b 𝐤 2 - 2 Im ( 𝐲 * μ 𝐟 𝐤 𝐟 𝐤 * μ * 𝐠 ) ,

which implies (3.22).

B Proof of Theorem 1

Denote

= span { 𝐆 - 𝒫 T ~ ( π ( 𝐆 ) ) 𝐆 }

and

= ( 𝒩 ρ ) [ T ( π ( 𝐆 ) ) ] .

We have the following orthogonal decompositions for 𝐆 and 𝒫 𝐆 .

Decomposition for G :

𝐆 = π ( 𝐆 ) + D 𝐆 + E 𝐆 ,

where D 𝐆 = 𝒫 T ~ ( π ( 𝐆 ) ) 𝐆 - π ( 𝐆 ) T ( π ( 𝐆 ) ) and E 𝐆 = 𝐆 - 𝒫 T ~ ( π ( 𝐆 ) ) 𝐆 .

Decomposition for P M G :

𝒫 𝐆 = π ( 𝐆 ) + D + E + F

for some D T ( π ( 𝐆 ) ) , E and F .

The orthogonality of T ( π ( 𝐆 ) ) , and implies

𝒫 T ~ ( π ( 𝐆 ) ) 𝐆 - 𝒫 𝐆 2 = π ( 𝐆 ) + D 𝐆 - ( π ( 𝐆 ) + D + E + F ) 2
(B.1) = D - D 𝐆 2 + E 2 + F 2 ,

where the decomposition ( D , E , F ) has the following constraint.

Lemma 1.

For any G C ( N ρ ) , a necessary condition for ( D , E , F ) to be the orthogonal decomposition of P M G - π ( G ) is

($C_{DEF}$) D - D 𝐆 2 + E - E 𝐆 2 + F 2 [ E 𝐆 + 1 2 𝐆 - π ( 𝐆 ) 2 ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 ] 2 .

The proof of Lemma 1 can be found in Appendix C. Because ($C_{DEF}$) is the necessary condition for ( D , E , F ) to be the orthogonal decomposition of 𝒫 𝐆 - π ( 𝐆 ) , (B.1) implies that

(B.2) 𝒫 T ~ ( π ( 𝐆 ) ) 𝐆 - 𝒫 𝐆 2 max C D E F D - D 𝐆 2 + E 2 + F 2 .

By the method of Lagrange multipliers, the maximum in (B.2) is attained when F = 0 , so (B.2) implies

(B.3) 𝒫 T ~ ( π ( 𝐆 ) ) 𝐆 - 𝒫 𝐆 2 max C D E D - D 𝐆 2 + E 2 ,

where ($C_{DE}$) is a constraint obtained by substituting F = 0 into the constraint ($C_{DEF}$), i.e.,

($C_{DE}$) D - D 𝐆 2 + E - E 𝐆 2 [ E 𝐆 + 1 2 𝐆 - π ( 𝐆 ) 2 ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 ] 2 .

Figure 2 
                  Constraint ($C_{DE}$) on D and E (
                        
                           
                              
                                 
                                    𝒫
                                    ℳ
                                 
                                 ⁢
                                 𝐠
                              
                           
                           
                           {\mathcal{P}_{\mathcal{M}}\mathbf{g}}
                        
                      is
contained in the shadow area and 
                        
                           
                              
                                 △
                                 =
                                 
                                    
                                       1
                                       2
                                    
                                    ⁢
                                    
                                       
                                          ∥
                                          
                                             𝐆
                                             -
                                             
                                                π
                                                ⁢
                                                
                                                   (
                                                   𝐆
                                                   )
                                                
                                             
                                          
                                          ∥
                                       
                                       2
                                    
                                    ⁢
                                    
                                       
                                          (
                                          
                                             
                                                ∑
                                                
                                                   𝐤
                                                   ∈
                                                   
                                                      𝒦
                                                      ρ
                                                   
                                                
                                             
                                             
                                                b
                                                𝐤
                                                
                                                   -
                                                   2
                                                
                                             
                                          
                                          )
                                       
                                       
                                          1
                                          2
                                       
                                    
                                 
                              
                           
                           
                           {\triangle=\frac{1}{2}\|\mathbf{G}-\pi(\mathbf{G})\|^{2}(\sum_{\mathbf{k}\in%
\mathcal{K}_{\rho}}b_{\mathbf{k}}^{-2})^{\frac{1}{2}}}
                        
                     ).
Figure 2

Constraint ($C_{DE}$) on D and E ( 𝒫 𝐠 is contained in the shadow area and = 1 2 𝐆 - π ( 𝐆 ) 2 ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 ).

The constraint ($C_{DE}$) is depicted in Figure 2, which implies that the right-hand side of (B.3) can be rewritten as

max C D E D - D 𝐆 2 + E 2 = [ 2 E 𝐆 + 1 2 𝐆 - π ( 𝐆 ) 2 ( 𝐤 𝒦 ρ b k - 2 ) 1 2 ] 2 .

In the following, we derive another constraint on E , which is motivated from [1, Proposition 2.4]. First of all, by the orthogonality property T ~ ( π ( 𝐆 ) ) ( ) , we have

distance ( 𝒫 𝐆 , T ~ ( π ( 𝐆 ) ) ) = distance ( π ( 𝐆 ) + D + E + F , T ~ ( π ( 𝐆 ) ) )
(B.4) = ( E 2 + F 2 ) 1 2 ,

where distance ( 𝐔 , 𝐕 ) is defined by the Frobenius norm of 𝐔 - 𝐕 . Note that E 0 when 𝐆 - π ( 𝐆 ) 0 . To get an effective constraint on E through (B.4), we need a more precise estimate for distance ( 𝒫 𝐆 , T ~ ( π ( 𝐆 ) ) ) . Note that

(B.5) distance ( 𝒫 𝐆 , T ~ ( π ( 𝐆 ) ) ) = 𝒫 𝐆 - 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 𝐆 ) .

We apply (3.21) and (3.22) to write down the vectorization of 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 𝐆 ) as follows:

vec ( 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 ( 𝐆 ) ) ) = vec ( π ( 𝐆 ) ) + i 𝐤 𝒦 ρ θ 𝐤 μ 𝐟 𝐤 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) ,

where

θ 𝐤 = Im ( vec ( π ( 𝐆 ) ) * μ 𝐟 k 𝐟 k * μ * vec ( 𝒫 𝐆 ) ) b 𝐤 2 .

Because

vec ( π ( 𝐆 ) ) * μ 𝐟 𝐤 𝐟 𝐤 * μ * vec ( 𝒫 𝐆 ) = 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) ¯ ( b 𝐤 𝐟 𝐤 * μ * 𝐠 | 𝐟 k * μ * 𝐠 | ) ,

we have

θ 𝐤 = Im ( 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) ¯ b 𝐤 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | ) b 𝐤 - 2 = Im ( 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) ¯ b 𝐤 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | ) .

Therefore,

(B.6) vec ( 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 ( 𝐆 ) ) ) - vec ( π ( 𝐆 ) ) = i k 𝒦 ρ Im ( 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) ¯ b 𝐤 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | ) μ 𝐟 𝐤 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) .

We start to estimate 𝒫 ( 𝐆 ) - 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 𝐆 ) (i.e., the right-hand side of (B.5)) as follows. By (3.9) and (B.6),

vec ( 𝒫 ( 𝐆 ) - 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 ( 𝐆 ) ) )
= μ k 𝒦 ρ 𝐟 𝐤 b 𝐤 𝐟 𝐤 * μ * 𝐠 | 𝐟 k * μ * 𝐠 | - μ k 𝒦 ρ 𝐟 𝐤 b 𝐤 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) b 𝐤 - i k 𝒦 ρ Im ( 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) ¯ b k 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | ) μ 𝐟 𝐤 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) )
(B.7) = μ 𝐤 𝒦 ρ 𝐟 𝐤 b 𝐤 { 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | - 𝐟 k * μ * vec ( π ( 𝐆 ) ) b 𝐤 - i Im ( 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) ¯ b 𝐤 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | ) 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) b 𝐤 } .

Because { 𝐟 𝐤 } 𝐤 𝒦 ρ are orthonormal vectors, (B.7) implies

𝒫 ( 𝐆 ) - 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 ( 𝐆 ) ) 2
(B.8) = 𝐤 𝒦 ρ b 𝐤 2 | 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | - 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) b 𝐤 - i Im ( 𝐟 k * μ * vec ( π ( 𝐆 ) ) ¯ b 𝐤 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | ) 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) b 𝐤 | 2 .

For every 𝐤 𝒦 ρ , we define ν k , ω k ( - π , π ] as follows:

e i ν k = 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | , e i ω k = 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) b 𝐤 ,

and define the difference | ν 𝐤 - ω 𝐤 | between ν 𝐤 and ω 𝐤 by

| ν 𝐤 - ω 𝐤 | = min { 2 π - | ν 𝐤 - ω 𝐤 | , | ν 𝐤 - ω 𝐤 | } .

The notation above helps us get the following estimation:

| 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | - 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) b 𝐤 - i Im ( 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) ¯ b 𝐤 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | ) 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) b 𝐤 | 2
= | e i ν 𝐤 - e i ω 𝐤 - i Im ( e - i ω 𝐤 e i ν 𝐤 ) e i ω 𝐤 | 2
= | e i ( ν 𝐤 - ω 𝐤 ) - 1 - i Im ( e i ( ν 𝐤 - ω 𝐤 ) ) | 2
= | cos ( ν 𝐤 - ω 𝐤 ) - 1 | 2
= | cos ( min { | ν 𝐤 - ω 𝐤 | , 2 π - | ν 𝐤 - ω 𝐤 | } ) - 1 | 2
= | cos ( | ν 𝐤 - ω 𝐤 | * ) - 1 | 2
(B.9) 1 4 | ν 𝐤 - ω 𝐤 | * 4 ,

where the inequality above holds for any ν k and ω k . By substituting (B.9) into (B), we get

(B.10) 𝒫 𝐠 - 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 𝐆 ) 2 = 1 4 𝐤 𝒦 ρ b 𝐤 2 | ν 𝐤 - ω 𝐤 | * 4 .

By the inequality

| ν 𝐤 - ω 𝐤 | 2 | e i ν 𝐤 - e i ω 𝐤 | , 𝐤 𝒦 ρ ,

(B.10) implies that

𝒫 𝐆 - 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 𝐆 ) 2 4 𝐤 𝒦 ρ b 𝐤 2 | e i ν 𝐤 - e i ω 𝐤 | 4
(B.11) 4 ( max 𝐤 𝒦 ρ b 𝐤 - 2 ) ( 𝐤 𝒦 ρ b 𝐤 2 | e i ν 𝐤 - e i ω 𝐤 | 2 ) 2 .

We claim that

(B.12) k 𝒦 ρ b 𝐤 2 | e i ν 𝐤 - e i ω 𝐤 | 2 4 𝐆 - π ( 𝐆 ) 2 ,

whose proof can be found in Appendix D. By substituting (B.12) into (B.11),

𝒫 𝐆 - 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 𝐆 ) 2 4 ( max 𝐤 𝒦 ρ b 𝐤 - 2 ) ( 4 𝐆 - π ( 𝐆 ) 2 ) 2 = 2 6 ( max 𝐤 𝒦 ρ b 𝐤 - 2 ) 𝐆 - π ( 𝐆 ) 4 .

By denoting β = 2 3 ( max 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 , we have

𝒫 𝐆 - 𝒫 T ~ ( π ( 𝐆 ) ) ( 𝒫 ( 𝐆 ) ) β 𝐆 - π ( 𝐆 ) 2 .

By combining the estimation above with (B.4) and (B.5), we obtain another constraint as follows:

(${C_{E}}$) E β 𝐆 - π ( 𝐆 ) 2 .

The entire constraints, including ($C_{DE}$) and (${C_{E}}$), on ( D , E ) are depicted in Figure 3.

Figure 3

Constraints ($C_{DE}$) and (${C_{E}}$) on D and E ( 𝒫 𝐆 belongs to the intersection of the shadow area andthe region { ( D , E ) : E β 𝐆 - π ( 𝐆 ) 2 } ).

(a)
(a)
(b)
(b)

Finally, we consider two cases as follows:

Case 1: 2 E G + 1 2 G - π ( G ) 2 ( k K ρ b k - 2 ) 1 2 β G - π ( G ) 2 . Using the notation described in Figure 3,

(B.13) 𝒫 𝐆 - 𝒫 T ~ ( π ( 𝐆 ) ) 𝐆 = 𝒫 𝐆 - ( π ( 𝐆 ) + D 𝐆 ) r max ,

where r max = r 1 2 + r 2 2 and r 1 = β 𝐆 - π ( 𝐆 ) 2 . Both for Figures 3 (a) and 3 (b),

r 2 2 = ( E 𝐆 + ) 2 - ( E 𝐆 - r 1 ) 2
= [ E 𝐆 + 1 2 𝐆 - π ( 𝐆 ) 2 ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 ] 2 - ( E 𝐆 - β 𝐆 - π ( 𝐆 ) 2 ) 2
= [ 2 E 𝐆 + ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 - 2 β 2 𝐆 - π ( 𝐆 ) 2 ] [ ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 + 2 β 2 ] 𝐆 - π ( 𝐆 ) 2
(B.14) 𝒪 1 ( 𝐆 - π ( 𝐆 ) ) 𝐆 - π ( 𝐆 ) 2 ,

where the last inequality follows from E 𝐆 𝐆 - π ( 𝐆 ) and

𝒪 1 ( 𝐆 - π ( 𝐆 ) ) = { 2 + [ 1 2 ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 - β ] + 𝐆 - π ( 𝐆 ) } [ 1 2 ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 + β ] 𝐆 - π ( 𝐆 ) .

By (B.14), we obtain

r max = r 1 2 + r 2 2
β 2 𝐆 - π ( 𝐆 ) 4 + 𝒪 1 ( 𝐆 - π ( 𝐆 ) ) 𝐆 - π ( 𝐆 ) 2
= β 2 𝐆 - π ( 𝐆 ) 2 + 𝒪 1 ( 𝐆 - π ( 𝐆 ) ) 𝐆 - π ( 𝐆 )
(B.15) = : 𝒪 2 ( 𝐆 - π ( 𝐆 ) ) 𝐆 - π ( 𝐆 ) .

The statement of Theorem 1 follows by substituting (B.15) into (B.13).

Case 2: 2 E G + 1 2 G - π ( G ) 2 ( k K ρ b k - 2 ) 1 2 < β G - π ( G ) 2 . This case implies that the constraint C E cannot enhance the constraint ($C_{DE}$) (i.e., the dashed horizontal lines in Figures 3 (a) and 3 (b) do not have any intersection with the shadow areas), so we directly get

𝒫 𝐆 - 𝒫 T ~ ( π ( 𝐆 ) ) 𝐆 2 E 𝐆 + 1 2 𝐆 - π ( 𝐆 ) 2 ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2
(B.16) < β 𝐆 - π ( 𝐆 ) 2

from the constraint ($C_{DE}$). Note that (B.16) implies the same conclusion as (B.15) by setting O 1 ( 𝐆 - π ( 𝐆 ) ) = 0 .

C Proof of Lemma 1

First of all, we show that Lemma 1 follows from the estimate

(C.1) distance ( π ( 𝐆 ) + D 𝐆 , ) 1 2 𝐆 - π ( 𝐆 ) 2 ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2

for any 𝐆 ( 𝒩 ρ ) . Because 𝐆 - 𝒫 𝐆 is the shortest distance from 𝐆 to , we have the triangle inequality

𝐆 - 𝒫 ( 𝐆 ) distance ( 𝐆 , π ( 𝐆 ) + D 𝐆 ) + distance ( π ( 𝐆 ) + D 𝐆 , )
(C.2) E 𝐆 + 1 2 𝐆 - π ( 𝐆 ) 2 ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 ,

where the last inequality is implied by (C.1). On the other hand, by the orthogonal decompositions of 𝐆 and 𝒫 𝐆 , we have

𝐆 - 𝒫 𝐆 2 = π ( 𝐆 ) + D 𝐆 + E 𝐆 - ( π ( 𝐆 ) + D + E + F ) 2
(C.3) = D - D 𝐆 2 + E - E 𝐆 2 + F 2 .

By combining (C.2) and (C.3), we get the desired result

D - D 𝐆 2 + E - E 𝐆 2 + F 2 [ E 𝐆 + 1 2 𝐆 - π ( 𝐆 ) 2 ( 𝐤 𝒦 ρ b 𝐤 - 2 ) 1 2 ] 2 .

Now we start to prove (C.1). Denote

π ( 𝐆 ) = { 𝐌 ( 𝒩 ρ ) : vec ( 𝐌 ) = μ Φ e i Θ Φ * μ * vec ( π ( 𝐆 ) ) , Θ  is a real-valued diagonal matrix of size  | 𝒦 ρ | } ,

where Φ is a matrix consisting of the column vectors { 𝐟 𝐤 } 𝐤 𝒦 ρ . By (3.18), we know that π ( 𝐆 ) , which implies

(C.4) distance ( π ( 𝐆 ) + D 𝐆 , ) distance ( π ( 𝐆 ) + D 𝐆 , π ( 𝐆 ) ) .

By denoting k = k 1 ( [ ρ n 1 ] + 1 ) 2 + k 2 ( [ ρ n 2 ] + 1 ) for any 𝐤 = ( k 1 , k 2 ) 𝒦 ρ , 𝐟 k = 𝐟 𝐤 , b k = b 𝐤 , and 𝐠 = vec ( 𝐆 ) , (3.21) and (3.22) imply that the vectorization of D 𝐆 can be expressed as

vec ( D 𝐆 ) = i μ Φ Λ 𝐆 Φ * μ * vec ( π ( 𝐆 ) ) ,

where

Λ 𝐆 = diag ( Im ( vec ( π ( 𝐆 ) ) * μ 𝐟 0 𝐟 0 * μ * 𝐠 ) b 0 2 , , Im ( vec ( π ( 𝐆 ) ) * μ 𝐟 | K ρ | - 1 𝐟 | K ρ | - 1 * μ * 𝐠 ) b | K ρ | - 1 2 ) .

Hence, the term on the right-hand side of (C.4) can be calculated explicitly as follows:

distance ( π ( 𝐆 ) + D 𝐆 , π ( 𝐆 ) ) = min Θ vec ( π ( 𝐆 ) ) + i μ Φ Λ 𝐆 Φ * μ * vec ( π ( 𝐆 ) ) - μ Φ e i Θ Φ * μ * vec ( π ( 𝐆 ) )
(C.5) = min Θ [ 𝐈 + i μ Φ Λ 𝐆 Φ * μ * - μ Φ e i Θ Φ * μ * ] vec ( π ( 𝐆 ) ) ,

where the minimum is taken over all diagonal matrices in | 𝒦 ρ | × | 𝒦 ρ | . Because μ Φ is unitary,

[ 𝐈 + i μ Φ Λ 𝐆 Φ * μ * - μ Φ e i Θ Φ * μ * ] vec ( π ( 𝐆 ) ) = μ Φ ( 𝐈 + i Λ 𝐆 - e i Θ ) Φ * μ * vec ( π ( 𝐆 ) )
= ( 𝐈 + i Λ 𝐆 - e i Θ ) Φ * μ * vec ( π ( 𝐆 ) ) ,

which implies that

min Θ [ 𝐈 + i μ Φ Λ 𝐆 Φ * μ * - μ Φ e i Θ Φ * μ * ] vec ( π ( 𝐆 ) ) 2
= k = 0 | 𝒦 ρ | - 1 min θ k | 1 + i Im ( vec ( π ( 𝐆 ) ) * μ 𝐟 k 𝐟 k * μ * 𝐠 ) b k 2 - e i θ k | 2 | [ Φ * μ * vec ( π ( 𝐆 ) ) ] k | 2
(C.6) = k = 0 | 𝒦 ρ | - 1 | [ 1 + ( Im ( vec ( π ( 𝐆 ) ) * μ 𝐟 k 𝐟 k * μ * 𝐠 ) b k 2 ) 2 ] 1 2 - 1 | 2 | [ Φ * μ * vec ( π ( 𝐆 ) ) ] k | 2 .

Applying the inequality ( 1 + δ ) 1 2 - 1 1 2 δ (for all δ 0 ) into (C.6), we get

min Θ [ 𝐈 + i μ Φ Λ 𝐆 Φ * μ * - μ Φ e i Θ Φ * μ * ] vec ( π ( 𝐆 ) ) 2
1 4 k = 0 | 𝒦 ρ | - 1 ( Im ( vec ( π ( 𝐆 ) ) * μ 𝐟 k 𝐟 k * μ * 𝐠 ) b k 2 ) 4 | [ Φ * μ * vec ( π ( 𝐆 ) ) ] k | 2
= 1 4 k = 0 | 𝒦 ρ | - 1 [ Im ( vec ( π ( 𝐆 ) ) * μ 𝐟 k 𝐟 k * μ * 𝐠 ) ] 4 b k 6
(C.7) 1 4 𝐆 - π ( 𝐆 ) 4 k = 0 | 𝒦 ρ | - 1 b k - 2 .

The claim (C.1) follows by (C.4), (C.5) and (C.7).

D Proof of (B.12)

Because π ( 𝐆 ) implies that b 𝐤 = | 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | for all 𝐤 𝒦 ρ , the claim (B.12) is equivalent to

(D.1) 𝐤 𝒦 ρ | 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | 2 | 𝐟 k * μ * 𝐠 | 𝐟 k * μ * 𝐠 | - 𝐟 k * μ * vec ( π ( 𝐆 ) ) | 𝐟 k * μ * vec ( π ( 𝐆 ) ) | | 2 4 𝐆 - π ( 𝐆 ) 2 .

Let

𝐕 = [ V 1 V 2 ] = [ vec ( π ( 𝐆 ) ) vec ( π ( 𝐆 ) ) ] | 𝒦 ρ | × | 𝒦 ρ |

be an orthonormal matrix such that 𝐕 * vec ( π ( 𝐆 ) ) vec ( π ( 𝐆 ) ) = 𝐞 1 and 𝐕 * 𝐠 = κ 1 𝐞 1 + κ 2 𝐞 2 , where 𝐞 1 and 𝐞 2 are the first and second columns in the standard basis of the real vector space | 𝒦 ρ | , and κ 1 , κ 2 . We have

𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | - 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | = 𝐟 𝐤 * μ * 𝐕𝐕 * 𝐠 | 𝐟 𝐤 * μ * 𝐕𝐕 * 𝐠 | - 𝐟 𝐤 * μ * 𝐕𝐕 * vec ( π ( 𝐆 ) ) | 𝐟 𝐤 * μ * 𝐕𝐕 * vec ( π ( 𝐆 ) ) |
(D.2) = 𝐟 𝐤 * μ * 𝐕 ( κ 1 𝐞 1 + κ 2 𝐞 2 ) | 𝐟 𝐤 * μ * 𝐕 ( κ 1 𝐞 1 + κ 2 𝐞 2 ) | - 𝐟 k * μ * 𝐕𝐞 1 | 𝐟 k * μ * 𝐕𝐞 1 | .

Denote 𝐟 𝐤 * μ * 𝐕 = [ r 𝐤 , 1 r 𝐤 , 2 r 𝐤 , | 𝒦 ρ | ] 1 × | 𝒦 ρ | . Then (D.2) implies that

| 𝐟 k * μ * 𝐠 | 𝐟 k * μ * 𝐠 | - 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | | = | κ 1 r 𝐤 , 1 + κ 2 r 𝐤 , 2 | κ 1 r 𝐤 , 1 + κ 2 r 𝐤 , 2 | - r 𝐤 , 1 | r 𝐤 , 1 | |
= | κ 1 r 𝐤 , 1 + κ 2 r 𝐤 , 2 | κ 1 r 𝐤 , 1 + κ 2 r 𝐤 , 2 | r 𝐤 , 1 ¯ | r 𝐤 , 1 | - 1 |
= | Phase ( κ 1 | r 𝐤 , 1 | 2 + κ 2 r 𝐤 , 1 ¯ r 𝐤 , 2 ) - 1 |
(D.3) = | Phase ( κ 1 π ( 𝐆 ) + κ 2 π ( 𝐆 ) r 𝐤 , 2 r 𝐤 , 1 ) - 1 | ,

where Phase ( y ) = y | y | for y { 0 } . For the function Phase ( ) , we have the following inequalities (see, for example, [26, Lemma A.7]):

| Phase ( 1 + y ) - 1 | = | Phase ( 1 + y ) - ( 1 + y ) + y |
min ( 2 , | Phase ( 1 + y ) - ( 1 + y ) | + | y | )
(D.4) min ( 2 , | | 1 + y | - 1 | + | y | ) 2 min ( 1 , | y | ) .

By applying the inequality (D.4) with y = κ 1 π ( 𝐆 ) + κ 2 π ( 𝐆 ) r 𝐤 , 2 r 𝐤 , 1 - 1 into (D.3), we get

(D.5) | 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | - 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | | 2 | κ 1 π ( 𝐆 ) + κ 2 π ( 𝐆 ) r 𝐤 , 2 r 𝐤 , 1 - 1 | .

Because | 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | = π ( 𝐆 ) | r 𝐤 , 1 | , (D.5) implies that

𝐤 𝒦 ρ | 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | 2 | 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | - 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | | 2
𝐤 𝒦 ρ ( π ( 𝐆 ) | r 𝐤 , 1 | ) 2 ( 2 | κ 1 π ( 𝐆 ) + κ 2 π ( 𝐆 ) r 𝐤 , 2 r 𝐤 , 1 - 1 | ) 2
= 4 𝐤 𝒦 ρ | r 𝐤 , 1 ( κ 1 - π ( 𝐆 ) ) + κ 2 r 𝐤 , 2 | 2
(D.6) = 4 { 𝐤 𝒦 ρ | r 𝐤 , 1 | 2 | κ 1 - π ( 𝐆 ) | 2 + | κ 2 | 2 | r 𝐤 , 2 | 2 + 2 Re [ κ 2 ¯ r 𝐤 , 2 ¯ r 𝐤 , 1 ( κ 1 - π ( 𝐆 ) ) ] } .

Because 𝐕 = [ V 1 V 2 V | 𝒦 ρ | ] is an orthonormal matrix and 𝐤 𝒦 ρ 𝐟 𝐤 𝐟 𝐤 * = 𝐈 | 𝒦 ρ | × | 𝒦 ρ | , we have

𝐤 𝒦 ρ r 𝐤 , 2 ¯ r 𝐤 , 1 = 𝐤 𝒦 ρ [ 𝐟 𝐤 * μ * V 2 ] * [ 𝐟 𝐤 * μ * V 1 ] = 0 ,

which implies that

(D.7) 𝐤 𝒦 ρ Re [ κ 2 ¯ r 𝐤 , 2 ¯ r 𝐤 , 1 ( κ 1 - π ( 𝐆 ) ) ] = 0 .

In the same way, we have

(D.8) 𝐤 𝒦 ρ | r 𝐤 , 1 | 2 = 𝐤 𝒦 ρ | r 𝐤 , 2 | 2 = 1 .

By substituting (D.7) and (D.8) into (D.6), we get

(D.9) 𝐤 𝒦 ρ | 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | 2 | 𝐟 𝐤 * μ * 𝐠 | 𝐟 𝐤 * μ * 𝐠 | - 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | 𝐟 𝐤 * μ * vec ( π ( 𝐆 ) ) | | 2 4 ( | κ 1 - π ( 𝐆 ) | 2 + | κ 2 | 2 ) .

The claim (D.1) follows by (D.9) and noticing that | κ 1 - π ( 𝐆 ) | 2 + | κ 2 | 2 = 𝐆 - π ( 𝐆 ) 2 .

Acknowledgements

The author gratefully acknowledges the anonymous reviewers for their constructive comments and critiques, which significantly enhanced this manuscript. The author also wishes to thank Albert Fannjiang and Pengwen Chen for their valuable discussions and insightful feedback.

References

[1] F. Andersson and M. Carlsson, Alternating projections on nontangential manifolds, Constr. Approx. 38 (2013), no. 3, 489–525. 10.1007/s00365-013-9213-3Search in Google Scholar

[2] A. S. Bandeira, Y. Chen and D. G. Mixon, Phase retrieval from power spectra of masked signals, Inf. Inference 3 (2014), no. 2, 83–102. 10.1093/imaiai/iau002Search in Google Scholar

[3] H. H. Bauschke, J. Y. Bello Cruz, T. T. A. Nghia, H. M. Phan and X. Wang, The rate of linear convergence of the Douglas–Rachford algorithm for subspaces is the cosine of the Friedrichs angle, J. Approx. Theory 185 (2014), 63–79. 10.1016/j.jat.2014.06.002Search in Google Scholar

[4] H. H. Bauschke, P. L. Combettes and D. R. Luke, Phase retrieval, error reduction algorithm, and Fienup variants: A view from convex optimization, J. Opt. Soc. Amer. A 19 (2002), no. 7, 1334–1345. 10.1364/JOSAA.19.001334Search in Google Scholar

[5] E. J. Candès, X. Li and M. Soltanolkotabi, Phase retrieval from coded diffraction patterns, Appl. Comput. Harmon. Anal. 39 (2015), no. 2, 277–299. 10.1016/j.acha.2014.09.004Search in Google Scholar

[6] E. J. Candès, X. Li and M. Soltanolkotabi, Phase retrieval via Wirtinger flow: Theory and algorithms, IEEE Trans. Inform. Theory 61 (2015), no. 4, 1985–2007. 10.1109/TIT.2015.2399924Search in Google Scholar

[7] P. Chen and A. Fannjiang, Fourier phase retrieval with a single mask by Douglas–Rachford algorithms, Appl. Comput. Harmon. Anal. 44 (2018), no. 3, 665–699. 10.1016/j.acha.2016.07.003Search in Google Scholar PubMed PubMed Central

[8] P. Chen, A. Fannjiang and G.-R. Liu, Phase retrieval with one or two diffraction patterns by alternating projections with the null initialization, J. Fourier Anal. Appl. 24 (2018), no. 3, 719–758. 10.1007/s00041-017-9536-8Search in Google Scholar

[9] Y. Chen and E. J. Candès, Solving random quadratic systems of equations is nearly as easy as solving linear systems, Advances in Neural Information Processing Systems, Curran Associates, Red Hook (2015), 739–747. Search in Google Scholar

[10] F. H. Clarke, Y. S. Ledyaev, R. J. Stern and P. R. Wolenski, Nonsmooth Analysis and Control Theory, Grad. Texts in Math. 178, Springer, New York, 2008. Search in Google Scholar

[11] J. C. Dainty and J. R. Fienup, Phase retrieval and image reconstruction for astronomy, Image Recovery: Theory and Application, Academic Press, San Diego (1987), 231–275. Search in Google Scholar

[12] F. Deutsch, Best Approximation in Inner Product Spaces, CMS Books Math./Ouvrages Math. SMC 7, Springer, New York, 2001. 10.1007/978-1-4684-9298-9Search in Google Scholar

[13] Y. C. Eldar and S. Mendelson, Phase retrieval: Stability and recovery guarantees, Appl. Comput. Harmon. Anal. 36 (2014), no. 3, 473–494. 10.1016/j.acha.2013.08.003Search in Google Scholar

[14] A. Fannjiang, Absolute uniqueness of phase retrieval with random illumination, Inverse Problems 28 (2012), no. 7, Article ID 075008. 10.1088/0266-5611/28/7/075008Search in Google Scholar

[15] A. Fannjiang and W. Liao, Phase retrieval with random phase illumination, JOSA A 29 (2012), no. 9, 1847–1859. 10.1364/JOSAA.29.001847Search in Google Scholar PubMed

[16] A. Fannjiang and W. Liao, Fourier phasing with phase-uncertain mask, Inverse Problems 29 (2013), no. 12, Article ID 125001. 10.1088/0266-5611/29/12/125001Search in Google Scholar

[17] M. H. Hayes, The reconstruction of a multidimensional sequence from the phase or magnitude of its Fourier transform, IEEE Trans. Acoust. Speech Signal Process. 30 (1982), no. 2, 140–154. 10.1109/TASSP.1982.1163863Search in Google Scholar

[18] X. Jiang, S. Rajan and X. Liu, Wirtinger flow method with optimal stepsize for phase retrieval, IEEE Signal Process. Lett. 23 (2016), no. 11, 1627–1631. 10.1109/LSP.2016.2611940Search in Google Scholar

[19] S. Kayalar and H. L. Weinert, Error bounds for the method of alternating projections, Math. Control Signals Systems 1 (1988), no. 1, 43–59. 10.1007/BF02551235Search in Google Scholar

[20] R. Kolte and A. Ozgur, Phase retrieval via incremental truncated Wirtinger flow, preprint (2016), https://arxiv.org/abs/1606.03196. Search in Google Scholar

[21] A. S. Lewis, D. R. Luke and J. Malick, Local linear convergence for alternating and averaged nonconvex projections, Found. Comput. Math. 9 (2009), no. 4, 485–513. 10.1007/s10208-008-9036-ySearch in Google Scholar

[22] A. S. Lewis and J. Malick, Alternating projections on manifolds, Math. Oper. Res. 33 (2008), no. 1, 216–234. 10.1287/moor.1070.0291Search in Google Scholar

[23] J. Miao, P. Charalambous, J. Kirz and D. Sayre, Extending the methodology of X-ray crystallography to allow imaging of micrometre-sized non-crystalline specimens, Nature 400 (1999), no. 6742, 342–344. 10.1038/22498Search in Google Scholar

[24] R. P. Millane, Phase retrieval in crystallography and optics, JOSA A 7 (1990), no. 3, 394–411. 10.1364/JOSAA.7.000394Search in Google Scholar

[25] B. S. Mordukhovich, Variational Analysis and Generalized Differentiation. I, Grundlehren Math. Wiss. 330, Springer, Berlin, 2006. 10.1007/3-540-31246-3Search in Google Scholar

[26] P. Netrapalli, P. Jain and S. Sanghavi, Phase retrieval using alternating minimization, Advances in Neural Information Processing Systems 26, Curran Associates, Red Hook (2013), 2796–2804. Search in Google Scholar

[27] D. Noll and A. Rondepierre, On local convergence of the method of alternating projections, Found. Comput. Math. 16 (2016), no. 2, 425–455. 10.1007/s10208-015-9253-0Search in Google Scholar

[28] V. Pohl, F. Yang and H. Boche, Phase retrieval from low-rate samples, Sampl. Theory Signal Image Process. 14 (2015), no. 1, 71–99. 10.1007/BF03549588Search in Google Scholar

[29] R. Tyrrell Rockafellar and R. J.-B. Wets, Variational Analysis, Grundlehren Math. Wiss. 317, Springer, Berlin, 2009. Search in Google Scholar

[30] G. Wang, G. B. Giannakis and Y. C. Eldar, Solving systems of random quadratic equations via truncated amplitude flow, IEEE Trans. Inform. Theory 64 (2018), no. 2, 773–794. 10.1109/TIT.2017.2756858Search in Google Scholar

[31] K. Wei, Solving systems of phaseless equations via Kaczmarz methods: A proof of concept study, Inverse Problems 31 (2015), no. 12, Article ID 125008. 10.1088/0266-5611/31/12/125008Search in Google Scholar

[32] Z. Wen, C. Yang, X. Liu and S. Marchesini, Alternating direction methods for classical and ptychographic phase retrieval, Inverse Problems 28 (2012), no. 11, Article ID 115010. 10.1088/0266-5611/28/11/115010Search in Google Scholar

[33] P. Yin and J. Xin, PhaseLiftOff: An accurate and stable phase retrieval method based on difference of trace and Frobenius norms, Commun. Math. Sci. 13 (2015), no. 4, 1033–1049. 10.4310/CMS.2015.v13.n4.a10Search in Google Scholar

[34] H. Zhang, Y. Chi and Y. Liang, Provable non-convex phase retrieval with outliers: Median truncated Wirtinger flow, International Conference on Machine Learning, JMLR, New York (2016), 1022–1031. Search in Google Scholar

Received: 2019-03-10
Revised: 2021-01-22
Accepted: 2024-10-28
Published Online: 2024-11-15
Published in Print: 2025-04-01

© 2024 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 4.11.2025 from https://www.degruyterbrill.com/document/doi/10.1515/jiip-2019-0021/html
Scroll to top button