Home Decomposition and graphical correspondence analysis of checkerboard copulas
Article Open Access

Decomposition and graphical correspondence analysis of checkerboard copulas

  • Oliver Grothe and Jonas Rieger EMAIL logo
Published/Copyright: September 18, 2024
Become an author with De Gruyter Brill

Abstract

We analyze optimal low-rank approximations and correspondence analysis of the dependence structure given by arbitrary bivariate checkerboard copulas. Methodologically, we make use of the truncation of singular value decompositions of doubly stochastic matrices representing the copulas. The resulting (truncated) representations of the dependence structures are sparse, in particular, compared to the number of squares on the checkerboard. The additive structure of the decomposition carries through to statistical functionals of the copula, such as Kendall’s τ or Spearman’s ρ , and also motivates similarity measures for checkerboard copulas. We link our analysis to continuous decompositions of copula densities and copula-generating algorithms and discuss further general properties of the decomposition and its truncation. For example, truncated series might lack nonnegativity, and approximation errors increase for monotonicity-like copulas. We provide algorithms and extensions that account for and counteract these properties. The low-rank representation is illustrated for various copula examples, and some analytical results are derived. The resulting correspondence analysis profile plots are analyzed, providing graphical insights into the dependence structure implied by the copula. An illustration is provided with an empirical data set on fuel injector spray characteristics in jet engines.

1 Introduction

Copulas are a standard tool for modeling random vectors, as they separate marginal and dependence modeling. A copula contains information on the likelihood of joint occurrence of random variables on their intrinsic quantile scale. For two-dimensional vectors, the copula thus encodes a possibly large or infinite two-dimensional frequency table specifying the joint likelihood of the transformed random vector. If finite, square, and scaled appropriately, this table can be interpreted as a checkerboard copula [15,29]. The tables are generally large and may contain redundant information, and assessing the incorporated dependence information is not straightforward. We apply the well-known decomposition and dimensionality reduction techniques of high-dimensional data analysis to this table, thereby decomposing the copula. The decomposition opens a wide range of further analyses, for example, to compute and analyze copula characteristics, plot meaningful two-dimensional plots of the copula, or build simpler, reasonable approximations of complicated dependency structures. Through the low-rank approximation, one can drastically decrease the number of items to be stored compared to the full checkerboard matrix, i.e., the square of the lattice size.

Checkerboard copulas can be obtained from empirical data or, for example, by discretizing continuous copulas [16,27]. In either way, the copula frequency table is a doubly stochastic matrix. Taking the doubly stochastic matrix, we apply correspondence analysis methods that are mainly based on singular value decomposition (SVD).

Additive decompositions of copulas using variable-specific functions already exist in the literature, but only for continuous representations. Continuous decompositions are considered, for example, in Mesiar and Najjari [33] or Rodríguez-Lallena [40] for the generation of new copulas and in Cuadras [8] for the decomposition of copulas. The checkerboard case differs from existing approaches and yields different decompositions, as discussed in Section 3.3. Durrleman et al. [16] mentioned SVD of checkerboard copulas but did not go into detail, and Cuadras [9] considered discrete and continuous decompositions of general bivariate distributions. In contrast to these studies, we concentrate on the decomposition of doubly stochastic matrices that represent checkerboard copulas, allowing us to focus on the features of copulas. We provide formulas for important statistical functionals, including Spearman’s ρ , Kendall’s τ , and Pearson’s ϕ 2 . Through the Frobenius distance between the matrices, we express the similarity of two checkerboard copulas in terms of their ϕ 2 .

Using the standard kit of correspondence analysis has obstacles for some copulas. Copulas such as the comonotonicity copula are costly to represent in standard SVD, as the corresponding frequency matrix is similar to an identity matrix, having full rank and many equally large singular vectors. Thus, approximations by truncating the SVD series have slowly decaying errors. Therefore, we propose to use a monotonicity-anchored representation (MAR) (adapted from [18] and [24]), taking into account the independence and comonotonicity-like characteristics. This representation does not change the singular vectors for symmetric copulas but can considerably reduce the approximation error. Also, the obtained truncations are not necessarily valid checkerboard copulas, as negative values can occur. We provide an algorithm that yields the nearest valid copula for the Frobenius norm by generalizing an algorithm by Zass and Shashua [47] and thus maps the obtained truncated (not doubly stochastic) matrix to the nearest doubly stochastic matrix. While this article is focused on the Frobenius error norm, we remark on using the Hellinger distance in Appendix B.

The frequency table decomposition corresponds to a decomposition of the discretized copula probability distribution function (PDF). Section 2.6 links our analysis to continuous decompositions, as in the literature on copula generation and continuous copula decomposition, and to cumulative distribution function (CDF) decompositions. Through the decomposition, we motivate a decomposition of the Gaussian copula into transformed Hermite polynomials.

Thus, this article makes several contributions. We define the decomposition of checkerboard copulas and give extensions of the approach for comonotonicity-like copulas and non-copula truncations. We link the approach to important existing copula concepts such as dependence measures, similarities of copulas, and continuous decompositions of copulas. We derive characteristics of the graphs obtained by the approach and thus provide a new method of graphical copula representations. Finally, we apply the approach to theoretical copula families of various complexities and an empirical data example from the engineering context.

This article is structured as follows. Section 2 describes the approach, including the extensions for comonotonicity-like copulas, non-copula truncations, and the computation of statistical functionals. We analyze the difference between decomposed copulas and draw the connection between discrete (checkerboard) and continuous decompositions. We provide the resulting decompositions for the well-known copulas of different complexities and for symmetric and asymmetric dependencies in Section 3. We use the graphical tools of correspondence analysis to interpret the two-dimensional graphs of copulas and apply the graphical tools to an empirical checkerboard from data on the fuel injection spray characteristics of jet engines in Section 4. Section 5 concludes this article.

2 Checkerboard copula decomposition and its characteristics

This section examines the SVD and its truncation for checkerboard copulas, i.e., doubly stochastic matrices. We introduce some notations in Section 2.1 and then define the truncated decomposition, including an MAR that accounts for dependencies similar to comonotonicity in Section 2.2. To correct negative matrix elements in the truncated representation, Section 2.3 formulates an algorithm to approximate the truncation by a doubly stochastic matrix. Sections 2.4 and 2.5 derive statistical functionals and similarity measures using the decomposition. Section 2.6 links the decompositions of continuous copulas and their discretized counterparts.

2.1 Doubly stochastic matrices from bivariate copulas

Let X and Y be random variables with CDF F X , Y and marginal CDFs F X and F Y , respectively. Through the well-known theorem of Sklar [45] the multivariate CDF F X , Y ( x , y ) can be decomposed as

F X , Y ( x , y ) = C ( F X ( x ) , F Y ( y ) ) ,

whereby the copula C encodes the dependence structure of X and Y . The copula C can also be seen as CDF of F X ( X ) and F Y ( Y ) , and thus, has the properties of a multivariate CDF with uniform margins, provided that X and Y are continuous. While the copula is unique for continuous random variables, it is only uniquely identified on the image of F X and F Y , respectively, in the discrete case.

A checkerboard copula [29] is a special type of copula that assumes a uniform mass within the squares of an evenly spaced lattice I n × I n ( I n = { 0 , 1 n , , 1 } ). Checkerboard copulas can be computed from empirical data or by the discretization of continuous copulas. The discretization facilitates the (asymptotic) comparison of discrete and continuous characteristics.

Any continuous copula C defines a doubly stochastic matrix C n on the grid I n × I n by evaluating C on I n × I n , i.e., C C ( u , v ) ( u , v I n ) [27,31] and computing

(1) C i , j n = n C i n , j n C i 1 n , j n C i n , j 1 n + C i 1 n , j 1 n for i , j = 1 , , n .

The properties of C n follow from the copula properties of C :

  1. C n has nonnegative entries as the defining equation (1) coincides with a scaled version of the 2-volume of the copula, which is nonnegative.

  2. From C ( u , 1 ) = C ( 1 , u ) = u ( u [ 0 , 1 ] ) and C ( u , 0 ) = C ( 0 , u ) = 0 ( u [ 0 , 1 ] ) follows for j [ n ]

    i = 1 n C i , j n = i = 1 n n C i n , j n C i 1 n , j n C i n , j 1 n + C i 1 n , j 1 n = 1 .

    An analogous computation shows j = 1 n C i , j n = 1 ( i [ n ] ) , and thus, the row and column sums of C n are 1.

The matrix C n is by construction square, together with (1) and (2), a doubly stochastic matrix. The element C i , j n ( i , j [ n ] , where [ n ] { 1 , 2 , , n } ) corresponds to the density of the checkerboard copula

(2) c ˆ ( u , v ) = n C u n , v n n = n i = 1 n j = 1 n C i , j n 1 u i 1 n , i n 1 v j 1 n , j n ( u , v R ) ,

in the rectangle

R i , j i 1 n , i n × j 1 n , j n .

We denote by 1 the indicator function. Thus, C i , j n can be interpreted naturally as a table of the likelihood of occurrence in the copula domain. Integration over equation (2) yields a checkerboard approximation of the copula CDF C ˆ and the conditional CDF C ˆ u V = v . The discretizations approximate the copula C with increasing n , and every copula C is the limit of its discretizations for n [see 27, Theorem 1].

2.2 SVD and MAR

Having the table of likelihood of occurrence, C n , correspondence analysis can be applied to the matrix C n to analyze the structural properties of the matrix. Correspondence analysis uses the SVD to compute low-dimensional approximations of the matrix. To this end, the SVD is truncated, yielding the nearest matrix of the specified rank according to the Frobenius or the spectral norm [34]. In correspondence analysis, the matrix C n is usually centered, and some scaling is applied to rows and columns to account for the sum differences of the rows or columns [18]. In the case of C n , the centering step is implemented by subtracting the matrix Π n n 1 1 1 from C n , where 1 is the vector of ones of suitable dimension. We denote this by

A n = G ( C n ) C n Π n .

Note that 1 n 1 is a (left and right) singular vector of C n with singular value 1, whereby 1 is the largest singular value for doubly stochastic matrices [37], and thus, the rank of A n is at most n 1 . We denote the SVD of A n by

(3) A n = U S V , with U = ( u 1 , , u n ) , S = diag ( s 1 , , s n ) , V = ( v 1 , , v n ) ,

where U and V are the orthogonal matrices and the singular values s k are in [ 0 , 1 ] and are sorted in descending order as usual.

The decomposition in equation (3) may be truncated by using only the n * n largest singular values of s , and the corresponding first n * columns of U and V :

T n * ( A n ) U : , 1 : n * S 1 : n * , 1 : n * ( V : , 1 : n * ) ,

where we will use T n * ( ) as a truncation operator of the argument’s SVD in the following. The truncated T n * ( A n ) yields an approximation of C n by applying the inverse function of G , i.e.,

G 1 ( T n * ( A n ) ) = T n * ( A n ) + Π n .

The truncated SVD yields low-rank approximations with small errors for matrices with a few large and many small (or zero) singular values. We will show examples in Section 3. However, in the copula context, many copulas share characteristics with the comonotonicity copula, an identity matrix with singular value 1 with multiplicity n , and thus, high approximation errors for small-rank representations. To “remove” the comonotonicity copula characteristics before applying the SVD, we suggest transforming the matrix C n so that we account for high frequencies on the diagonal of the matrix and thus the monotone dependence structures. We denote this transformation by G MAR ( ) and call it MAR. As we argue in Lemma 1, through this representation, the singular vectors do not change for symmetric copulas, but the series of singular values s k decreases faster, leading to better low-rank approximations. The MAR is given by

(4) A ˜ n = G MAR ( C n , η ) c + η I n ( 1 + η ) 1 n 1 1 ,

with η R and I n denoting the n -by- n identity matrix. The centering step is implemented by the last summand ( 1 + η ) 1 n 1 1 , i.e., for η R ,

( A ˜ n ) 1 = C n 1 + η I n 1 1 + η n 1 1 1 = 1 + η 1 1 + η n 1 n = 0 1

and analogously for ( A ˜ n ) 1 . The approach also suits strong negative dependence by rotating the copula first. A similar transformation to G MAR can be found in Kazmierczak [24], in Greenacre [18, Section 8.6] formulated in the context of frequency tables. Unlike Greenacre [18] who used two parameters and demands them to be chosen such that ( A ˜ n ) i , j 0 ( i , j ) for merely illustrative purposes, we do not require this additional restriction here. We scale the last summand by 1 n to preserve the margins shown earlier.

Note that G ( C n ) is nested within the MAR by setting η = 0 . Later, the parameter η is calculated such that the Frobenius distance between the inverse transformed version of A ˜ n , denoted by G MAR 1 ( T n * ( A ˜ n ) , η ) , and (the original) A n is minimized.

Analogously to the aforementioned notation, we denote the SVD of A ˜ n by

A ˜ n = U ˜ S ˜ V ˜ , with U ˜ = ( u ˜ 1 , , u ˜ n ) , S ˜ = diag ( s ˜ 1 , , s ˜ n ) , and V ˜ = ( v ˜ 1 , , v ˜ n ) .

The following lemma shows that singular values and vectors of A ˜ n and A n are closely connected, provided that C n is symmetric.

Lemma 1

For the SVD of A ˜ n = U ˜ S ˜ V ˜ and A n = U S V of symmetric C n , i.e., for U = V , it holds that

u k = u ˜ k = v k = v ˜ k k = 1 , 2 , and s ˜ k = ( s k + η ) k = 1 , 2 , , n .

Proof

From v 1 , , v n 1 1 follows for k = 1 , , n 1

A ˜ n v k = C n + η I n + 1 + η n 1 1 v k = C n v k + η I n v k + 1 + η n 1 1 v k = 0 = s k u k + η v k .

For symmetric matrices, thus,

A ˜ n v k = ( s k + η ) v k .

For asymmetric C n , the singular values and vectors of A n and A ˜ n differ. Lemma 1 yields the n * -truncated representation of A ˜ n

T n * ( A ˜ n ) = k = 1 n * u ˜ k s ˜ k v ˜ k = ( * ) k = 1 n * ( s k + η ) u k u k ( * ) : for symmetric C n

and thus after backtransformation of equation (4)

(5) G MAR 1 ( T n * ( A ˜ n ) , η ) = k = 1 n * u ˜ k s ˜ k v ˜ k η I n + ( 1 + η ) 1 n 1 1

(6) = ( * ) k = 1 n * u k u k ( s k + η ) η I n + ( 1 + η ) 1 n 1 1

(7) = k = 1 n * u k u k ( s k + η ) η I n + ( 1 + η ) Π n

and equations (6) and (7) are, again, only valid for symmetric copulas. PDF and CDF can be computed using G MAR 1 ( T n * ( A ˜ n ) , η ) analogously to equation (2). The parameter η of equation (4) can be determined by minimizing some error norm of interest. For example, we calculate the fraction η that minimizes the residual inertia (thus, Frobenius error) for a specified approximation of rank n * by

(8) η * ( C n , n * ) = arg min η R k = 1 n * u ˜ k v ˜ k ( s ˜ k ) η I n + ( 1 + η ) Π n C n F 2 .

For a symmetric matrix C n and an approximation of dimension n * , this yields

(9) η * ( C n , n * ) = arg min η R k = n * + 1 n ( s k + η ) 2 = k = n * + 1 n s k n n * .

For asymmetric matrices C n , the problem in equation (8) can be solved numerically. The simulations in Section 3 examine the choices of η and the resulting matrices A ˜ n .

2.3 Ensuring double stochasticity of truncations

As noted earlier, truncations of the SVD can yield low error approximations with considerably lower rank matrices. In general, truncations of the SVD are not necessarily doubly stochastic matrices. Truncations keep the property of having row and column sums of one as the singular vectors u k and v k , or u ˜ k and v ˜ k , respectively, are perpendicular to 1 for k [ n 1 ] , but the truncations do not necessarily have nonnegative elements. One can approximate the truncation by the nearest, doubly stochastic matrix to ensure nonnegativity. This step does not increase the complexity of the representation, as it does not include any information other than the truncated matrix. We give a general idea of the algorithms for symmetric and asymmetric matrices here; they are more specifically described in Appendix A.

Zass and Shashua [47] proposed an algorithm to find the nearest doubly stochastic matrix for any symmetric matrix A sym { A R n × n : A = A } according to the Frobenius norm, i.e., a solution to the problem P ( A ) , with

P ( A ) = arg min B R n × n A B F 2 s.t. B 1 = 1 B 1 = 1 B i , j 0 , i , j [ n ] .

According to Zass and Shashua [47], P ( A ) can be solved for symmetric A iteratively by solving two problems, i.e., P ( A ) = P 2 P 1 P 2 P 1 ( A ) , with

P 1 ( A 1 ) = arg min B R n × n A 1 B F 2 s.t. B 1 = 1 B 1 = 1

and

P 2 ( A 2 ) = arg min B R n × n A 2 B F 2 s.t. B i , j 0 , i , j [ n ] .

A 1 and A 2 refer to the iterative solutions of P 2 and P 1 , respectively. Algorithm 1 formulates the algorithm explicitly. P 1 and P 2 have closed-form solutions and calculations for the solution of P 1 and P 2 are carried out in A. In the case of asymmetric matrices, Algorithm 1 retains its overall structure while incorporating a deflection component to maintain validity [17]. The resulting Algorithm 2 is shown in A. Note that there are algorithms for approximations with a particular interest in keeping the sparsity structure of A . Rontsis and Goulart [41] formulated an algorithm for a slightly modified problem that accounts for the sparsity of the matrix A based on the alternate direction method of multipliers and applied to symmetric and asymmetric matrices A . Sparsity thereby refers to zero entries in matrix A . In general, the SVD approximations typically contain many small, nonzero values, and thus, the approximation does not benefit from exploiting the sparsity structure.

2.4 Statistical functionals of decompositions and truncations

Various statistical properties can be computed using the decomposition, including dependence measures such as Kendall’s τ , Spearman’s ρ S , and Pearson’s ϕ 2 . We start by expressing the well-known dependency measures Kendall’s τ and Spearman’s ρ S through the decomposition. The structure of both measures inherits the SVD structure of the checkerboard copula, and thus, copulas with many high singular values tend to have a measure representation with many terms, subject to the direction of the singular vectors. The empirical computation of dependence measures using the checkerboard copula itself might be inefficient, and the use of a low-rank approximation might be more robust. We leave the asymptotics of the decomposed measures for further research.

Durrleman et al. [16] showed that for checkerboard copulas, Kendall’s τ and Spearman’s ρ S can be computed by

τ ( C n ) 1 1 n 2 trace ( E C n E ( C n ) ) and ρ S ( C n ) 3 n trace ( Ω C n ) 3 ,

with E R n × n and

E i , j = 1 , if i = j , 2 , if i > j , 0 , if i < j ,

and Ω R n × n , where

Ω ω ˇ ω ˇ , with ω ˇ = 1 n ( 2 n 2 1 + 1 , 2 n 2 2 + 1 , ) .

Let, as in Section 2.2, the SVD of the centered C n be denoted by A n = U S V , and additionally, u 0 = v 0 = 1 n 1 and s 0 = 1 , such that C n = k = 0 n 1 u k s k v k . Then, follows Spearman’s ρ S with ω ω ˇ 1 ω ˇ as

(10) ρ S ( C n ) = 3 n trace Ω k = 0 n 1 u k s k v k 3 = ( 4 1 n 2 ) k = 1 n 1 s k ω , u k v k , ω 3 = ( 4 1 n 2 ) k = 1 n 1 s k ω , u k v k , ω

and for Kendall’s τ

(11) τ ( C n ) 1 1 n 2 trace ( E C n E ( C n ) ) = 1 1 n 2 k = 0 n 1 l = 0 n 1 s k s l u l , E u k v k , E v l .

Details of the calculations are provided in Appendix C. Both dependence measures can also be put in terms of the MAR, for example,

(12) ρ S ( C n ) = 4 1 n 2 k = 1 n 1 s ˜ k u ˜ k , ω v ˜ k , ω 4 1 n 2 η + 3 n ( 1 + η ) 3

(13) = 4 1 n 2 k = 1 n 1 ( s k + η ) u ˜ k , ω 2 4 1 n 2 η + 3 n ( 1 + η ) 3 ( for symmetric C n ) .

Note that η refers to the MAR coefficient of equation (4). The calculations are performed in Appendix C. The decompositions of ρ S and τ are both based on the singular-value-weighted sum of scalar products containing the singular vectors. Thus, they account for the importance and the direction of the component. For τ , the projection vector contains transformations of the other singular vectors, reflecting the integral’s measure being the copula CDF. The representations in equations (10) and (11) yield approximations for Spearman’s ρ S and Kendall’s τ in terms of the truncated representations, i.e.,

(14) ρ S ˆ n * ( C n ) ( 4 1 n 2 ) k = 1 n * s k ω , u k v k , ω and

(15) τ ˆ n * ( C n ) 1 1 n 2 k = 0 n * l = 0 n * s k s l u l , E u k v k , E v l .

In the SVD representation, Pearson’s ϕ 2 boils down to the total inertia of the copula from independence [44, p. 223]

ϕ 2 ( C n ) = 0 1 0 1 c ˆ 2 ( u , v ) d u d v 1 = i = 1 n j = 1 n n 2 ( C i j n ) 2 1 n 2 1 = A n F 2 = k = 1 n s k 2 ,

where s 1 , s 2 , are the singular values of the centered A n . Note that this is proportional to Pearson’s χ 2 statistic for testing independence in an empirical contingency table (for Pearson’s χ 2 statistic in the copula context, see, e.g., [43]). The truncated representation ϕ 2 is

ϕ 2 ˆ n * ( C n ) = k = 1 n * s k 2 = ϕ 2 ( G 1 ( T n * ( A n ) ) ) .

In correspondence analysis, the ratio of the total inertia of approximation and the original matrix is a standard measure for the approximation’s goodness of fit, i.e.,

k = 1 n * s k 2 k = 1 n s k 2 = ϕ 2 ˆ n * ( C n ) ϕ 2 ( C n ) .

Counting the number of nonzero singular values yields an estimate of the dimensionality of the representation, i.e.,

ϕ g ( C n ) = { s k : s k > 0 , k [ n ] } = rank ( A n ) .

It counts the dimensions needed to model all information in C n and does not consider the strength of the information, in contrast to, for example, Pearson’s ϕ 2 . Cuadras and Díaz [12] called this the geometric dimension of a copula. For discretizations of a continuous copula, the values of ϕ 2 and ϕ g depend on the grid resolution n and are, therefore, the discretized copula’s properties and not of the continuous counterpart. The following lemma formulates this explicitly, and the example in Section 2.6 shows that the geometric dimension can decrease with increasing grid size when the grid sizes are not nested.

Lemma 2

Let n 1 > n 2 , with n 1 = n 2 m ( m N ), be the grid resolutions of the discretizations C n [ n 1 ] and C n [ n 2 ] of a copula C . Then,

ϕ g ( C n [ n 1 ] ) ϕ g ( C n [ n 2 ] ) .

Proof

Let M 1 and M 2 be the discretized copula C with grid sizes n 1 and n 2 , respectively. Then,

M 2 = A M 1 A ,

with A R n 2 × n 1 and A i j = 1 m 1 j ( ( i 1 ) m , i m ] and

rank ( M 2 ) = rank ( A M 1 A ) min ( rank ( A ) , rank ( M 1 ) , rank ( A ) ) = min ( rank ( A ) , rank ( M 1 ) ) rank ( M 1 ) .

2.5 Similarity of copulas

Using the decomposition makes it easy to compute the similarity of copulas if they have a shared grid size. We show that this similarity in terms of the Frobenius distance is mainly driven by Pearson’s ϕ 2 of the product of the two copulas. The Frobenius distance is highly dependent on the grid size n ; thus, we propose two normalizations. Let A A = U A S A ( V A ) and A B = U B S B ( V B ) be the two matrices after centering the bistochastic matrices C A and C B , respectively. For ease of notation, we omit the common grid size n . Then,

(16) A A A B F 2 = C A C B F 2 = trace C A ( C A ) 2 trace C B ( C A ) + trace C B ( C B ) = ϕ 2 ( C A ) + ϕ 2 ( C B ) 2 ϕ 2 ( C P ) = k = 1 n 1 ( s k A ) 2 + k = 1 n 1 ( s k B ) 2 2 trace A B ( A A ) + 2 ,

with the product copula C P = C B ( C A ) (see [27], p. 700). Whereas the terms ϕ 2 ( C A ) and ϕ 2 ( C B ) depend on the individual copulas solely, trace A B ( A A ) depends on the relative orientation of the singular vectors, i.e.,

trace A B ( A A ) = k = 1 n 1 l = 1 n 1 s k B s l A v k B , v l A u l A , u k B = k = 1 n 1 l = 1 n 1 s k B s l A cos α ( v k B , v l A ) cos α ( u l A , u k B ) ,

where α ( , ) is the angle between the two vectors. Thus, the copula similarity is driven by the similarity of the singular vectors weighted by the singular values.

Although the distance (squared) in equation (16) is straightforward to compute, it depends on the grid size n , as the range of values increases with n . Clearly, A A A B F 2 0 and A A A B F 2 = 0 for A A = A B . The maximum

(17) A A A B F 2 = ϕ 2 ( C A ) + ϕ 2 ( C B ) 2 ϕ 2 ( C P ) n + n 2 0 = 2 n

is attained, for example, for C A = I n and C B any doubly stochastic matrix with ones on off-diagonal elements, for example,

C B = 0 1 0 0 0 0 1 0 1 0 .

Thus, the use of the Frobenius distance suffers from a high dependence on the grid size n , and we propose two simple rescalings of the distance that account for the increase in n . The first one uses the maximal distance from equation (17), yielding

δ 1 ( C A , C B ) = C A C B F 2 n ,

so that the values lie within [ 0 , 1 ] . The examples in Section 3.4 indicate that this normalization overcorrects, resulting in decreasing δ 1 for checkerboard approximations of the same copulas with increasing n .

Another approach is to standardize the distance by the sum of Pearson’s ϕ 2 of the copulas C A and C B , i.e.,

δ 2 ( C A , C B ) = C A C B F ϕ 2 ( C A ) + ϕ 2 ( C B ) .

As C A C B F 0 , δ 2 ( , ) 0 and from C A C B F 2 ϕ 2 ( A A ) + ϕ 2 ( A B ) 2 ϕ 2 ( C P ) = 1 follows δ 2 ( , ) 1 . This standardization yields values that exhibit less variation with n . The similarity measures are applied to copulas in Section 3.4.

2.6 Some considerations on the link to continuous decompositions

Cuadras and Díaz [12] and Cuadras [8] defined continuous PDF decompositions for continuous copulas. In the following, we briefly expand on the connection between the continuous decomposition and the decomposition of the corresponding checkerboard copulas. Let again C denote the copula CDF, c the copula PDF and

(18) c ( u , v ) = 1 + k 1 λ k a k ( u ) b k ( v ) ,

with complete orthonormal sets { a k } and { b k } . Cuadras and Díaz [12] call the cardinality of the set { λ k : λ k 0 } geometric dimensionality, γ , of the copula, provided that γ is finite, analogously to the discretized case. The decomposition exists if the copula’s ϕ 2 , i.e., ϕ 2 ( C ) = k λ k 2 , is finite and induces a decomposition of the copula CDF

C ( u , v ) = 0 u 0 v 1 + k = 1 γ λ k a k ( u ¯ ) b k ( v ¯ ) d u ¯ d v ¯ = u v + k = 1 γ 0 u 0 v λ k a k ( u ¯ ) b k ( v ¯ ) d u ¯ d v ¯ .

The discretized copula of grid size n yields

C i j n = C i n , j n C i 1 n , j n C i n , j 1 n + C i 1 n , j 1 n = k = 1 γ λ k i 1 n i n a k ( u ) d u j 1 n j n b k ( v ) d v ,

and with the additional vectors

(19) a k = 0 1 n a k ( u ) d u , 1 n 2 n a k ( u ) d u , , ( k [ γ ] ) ,

(20) b k = 0 1 n b k ( v ) d v , 1 n 2 n b k ( v ) d v , , ( k [ γ ] ) ,

(21) C n = k = 1 γ λ k a k b k .

Note that equation (21) denotes an exact decomposition of C n , but not necessarily the SVD decomposition. The difference becomes particularly evident if n < γ and the summation in (21) has more summands than the dimensionality of the n × n matrix C n . To be the SVD, the vectors a k and b k must be left and right singular vectors. Take a l with l [ γ ] ,

(22) ( C n ) a l = k = 1 γ λ k a k b k a l

(23) = k = 1 γ λ k b k a k a l

(24) = k = 1 γ λ k b k ( a k a l ) .

Thus, a l is the left singular vector if it is orthogonal to the other { a k : k [ γ ] , k l } and if b l is a right singular vector. The corresponding singular value is λ k . The orthogonality condition is trivially fulfilled for γ = 1 and generally depends on the grid size n . The orthogonal { a k } and { b k } do not induce the orthogonality of the vectors { a k : k [ n ] } and { b k : k [ n ] } .

In addition, the decomposition in equation (21) bounds the geometric dimension of the discretized decomposition by the geometric dimension of the continuous decomposition. The trivial matrix-order bound is n 1 . Example 1 shows that this is, indeed, an upper bound and not an equality. A representation with fewer summands could be possible with fewer orthogonal vectors.

Example 1

Let C 1 be a continuous copula with uniform support on the rectangles shown in Figure 1(a). The continuous copula’s decomposition has geometric dimension γ ( C 1 ) = 3 . Figure 1(c) shows the geometric dimensions of discretizations of C 1 with various grid sizes. For n = 4 , the geometric dimension of the discretization is 1, and thus strictly smaller than the continuous geometric dimension and n 1 (Figure 1(b)).

Figure 1 
                  Example for a copula 
                        
                           
                           
                              
                                 
                                    C
                                 
                                 
                                    1
                                 
                              
                           
                           {C}_{1}
                        
                      with corresponding doubly stochastic matrix 
                        
                           
                           
                              
                                 
                                    C
                                 
                                 
                                    n
                                 
                              
                           
                           {{\bf{C}}}^{n}
                        
                      that has strictly smaller geometric dimension 
                        
                           
                           
                              γ
                              
                                 (
                                 
                                    
                                       
                                          C
                                       
                                       
                                          n
                                       
                                    
                                 
                                 )
                              
                           
                           \gamma \left({{\bf{C}}}^{n})
                        
                      than 
                        
                           
                           
                              min
                              
                                 (
                                 
                                    γ
                                    
                                       (
                                       
                                          C
                                       
                                       )
                                    
                                    ,
                                    n
                                    ‒
                                    1
                                 
                                 )
                              
                           
                           \min \left(\gamma \left(C),n&#x2012;1)
                        
                      for 
                        
                           
                           
                              n
                              =
                              4
                           
                           n=4
                        
                     : (a) the continuous copula’s mass is uniformly distributed over the blue rectangles, (b) the checkerboard copula’s mass is uniformly distributed over the blue rectangles for 
                        
                           
                           
                              n
                              =
                              4
                           
                           n=4
                        
                     , and (c) the geometric dimension 
                        
                           
                           
                              γ
                              
                                 (
                                 
                                    
                                       
                                          C
                                       
                                       
                                          n
                                       
                                    
                                 
                                 )
                              
                           
                           \gamma \left({{\bf{C}}}^{n})
                        
                      for increasing 
                        
                           
                           
                              n
                           
                           n
                        
                     .
Figure 1

Example for a copula C 1 with corresponding doubly stochastic matrix C n that has strictly smaller geometric dimension γ ( C n ) than min ( γ ( C ) , n 1 ) for n = 4 : (a) the continuous copula’s mass is uniformly distributed over the blue rectangles, (b) the checkerboard copula’s mass is uniformly distributed over the blue rectangles for n = 4 , and (c) the geometric dimension γ ( C n ) for increasing n .

Similar to the decompositions of the continuous copula, the decompositions of the copula CDF do not directly yield decompositions of the PDF. A continuous decomposition of the CDF with d summands is in general form

(25) C ( u , v ) = u v + k = 1 d λ k F k ( u ) G k ( v ) ,

with orthogonal F 1 , , F d and G 1 , , G d . It implies a decomposition of the PDF for differentiable F k and G k ( k = 1 , , d ),

c ( u , v ) = 2 C u v ( u , v ) = 1 + k = 1 d λ k F k u ( u ) G k v ( v ) ,

that generally lacks the orthogonality of the function F 1 ( u ) u , F 2 ( u ) u , . However, the aforementioned calculation shows that the number of summands for a representation of PDF is, at most, the number of summands of CDF, such that d is an upper bound for the geometric dimension, γ , of a PDF decomposition.

Equations (25) and (18) enable constructing copulas from appropriate { λ k } , { f k } , and { g k } . Rodríguez-Lallena [40] formulated conditions on the components to ensure the validity of the resulting copula. Mesiar and Najjari [33] extended this construction to higher (finite) dimensions. The construction only yields copulas without tail dependence and thus excludes, for example, the Cuadras-Augé (CA) or Gumbel copula. Instead of estimating all components, some parts in equation (25) can be fixed. Bakam and Pommeret [2], for example, used a Legendre polynomial basis and only fitted the remaining coefficients. In Section 3.3, we show that for (transformed) Hermite polynomials and certain { λ k } , the Gaussian copula arises. Allowing not only pairs in equation (18) but also the cross products for a k ( u ) b l ( v ) ( k l ) in the summation leads to further copula decomposition methods (see, for example, called generalized partition of unity copulas) [30,38].

We give further examples of the difference between continuous and discretized decomposition for the Farlie-Gumbel-Morgenstern (FGM) copula in Section 3.1 and for the Gaussian copula in Section 3.3.

3 Illustrative SVDs of copulas

This section provides the resulting decompositions for some checkerboard approximations of parametric copula families. Section 3.1 focuses on symmetric copulas, whereas Section 3.2 analyzes asymmetric copulas. These sections give examples of the resulting singular values and singular vectors, and we expand on the Frobenius norm-minimizing choice of η in the MAR. At the end of Section 3.1, we provide examples of invalid, i.e., non-copula, truncations, and the use of Algorithm 1. Section 3.3 compares the checkerboard and continuous decomposition, as introduced in Section 2.6, for the Gaussian copula. Section 3.4 applies the similarity measures of Section 2.5 to various checkerboard copulas.

In this section, we will denote the rank of the truncation by n * N 0 and refer to the non-MAR model by raw model.

3.1 Decompositions of symmetric copulas

We start with simple copulas with low geometric dimensions and obtain up to high geometric-dimensional copulas with tail dependence in the later examples in this section. The independence copula

C Π ( u , v ) = u v

yields the checkerboard copula C n = Π of geometric dimension 0. The comonotonicity copula

C M ( u , v ) = min ( u , v )

yields the checkerboard copula C n = I n with geometric dimension n 1 . The MAR with η = 1 fully recovers the matrix for n * 0 . Thus, the geometric dimensionality is significantly reduced in the MAR for the comonotonicity copula.

The FGM copula family with CDF

C θ , F G M ( u , v ) = u v + θ u v ( 1 u ) ( 1 v )

for θ [ 1 , 1 ] is of geometric dimensionality 1. Figure 2(a) depicts the first singular vector with respect to n and Figure 2(b) the first singular value, s 1 = θ 3 for the continuous representation with respect to θ . The first singular vector is u 1 = α ( 1 , 1 2 ( n 1 ) , , 1 ) ( α R such that u 2 = 1 ), being the checkerboard analog of a 1 ( x ) = 3 ( 1 2 x ) according to equation (19). The singular vector is the piecewise integrated a 1 since the geometric dimension is one. For n * = 1 , the MAR following the optimization in (9) has parameters η = s 1 ( n 1 ) = θ ( 3 ( n 1 ) ) . A numerical optimization in MATLAB yields numerically equivalent values, as shown in Figure 2(c). The approximation is improved with the MAR, but the gain is smaller than for the comonotonicity copula (Figure 2(d)). The matrix can be fully recovered for any n * 1 . The calculation of Spearman’s ρ S according to the representation in equation (10) yields the result for the FGM copula ( n * 1 )

ρ S ˆ ( C ˆ θ , F G M ) = ( 4 1 n 2 ) k = 1 n * s k ω , u k v k , ω = ( 4 1 n 2 ) θ 3 α 1 ω ˜ ω ˜ , ( 1 , 1 2 n , , 1 ) 2 = θ 3 ( 1 1 n 2 ) .

The approximated ρ S ˆ yields the FGM copula’s analytical Spearman’s ρ S of θ 3 for n .

Figure 2 
                  Analysis of the FGM checkerboard copula decompositions using the raw and MAR model: (a) elements 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                    ,
                                    j
                                 
                              
                           
                           {u}_{1,j}
                        
                      (
                        
                           
                           
                              j
                              ∈
                              
                                 [
                                 
                                    n
                                 
                                 ]
                              
                           
                           j\in \left[n]
                        
                     ) of the first singular vector 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                      for 
                        
                           
                           
                              θ
                              =
                              0.8
                           
                           \theta =0.8
                        
                      and various grid resolutions 
                        
                           
                           
                              n
                           
                           n
                        
                     . The same plots arise for other values of 
                        
                           
                           
                              θ
                              ≠
                              0
                           
                           \theta \ne 0
                        
                     . The different slopes result from the normalization of the singular vector, (b) the first singular value 
                        
                           
                           
                              
                                 
                                    s
                                 
                                 
                                    1
                                 
                              
                           
                           {s}_{1}
                        
                      for various values of 
                        
                           
                           
                              θ
                           
                           \theta 
                        
                     . The value is, by definition, positive, (c) the values of 
                        
                           
                           
                              η
                           
                           \eta 
                        
                      in the MAR minimizing the Frobenius error for a 0-truncation, which is only the MAR. The values for 
                        
                           
                           
                              η
                           
                           \eta 
                        
                      are obtained by numerical minimization using MATLAB’s fminsearch. The resulting values of 
                        
                           
                           
                              η
                           
                           \eta 
                        
                      coincide with their theoretical counterparts (see equation (9), and (d) Frobenius error of the MAR and the standard representation for 0-truncations. The values of 
                        
                           
                           
                              η
                           
                           \eta 
                        
                      are in plot c). The MAR reduces the error slightly.
Figure 2

Analysis of the FGM checkerboard copula decompositions using the raw and MAR model: (a) elements u 1 , j ( j [ n ] ) of the first singular vector u 1 for θ = 0.8 and various grid resolutions n . The same plots arise for other values of θ 0 . The different slopes result from the normalization of the singular vector, (b) the first singular value s 1 for various values of θ . The value is, by definition, positive, (c) the values of η in the MAR minimizing the Frobenius error for a 0-truncation, which is only the MAR. The values for η are obtained by numerical minimization using MATLAB’s fminsearch. The resulting values of η coincide with their theoretical counterparts (see equation (9), and (d) Frobenius error of the MAR and the standard representation for 0-truncations. The values of η are in plot c). The MAR reduces the error slightly.

The CA family of copulas [10] with CDF

C θ , CA ( u , v ) = u v 1 θ , u v , u 1 θ v , u v ,

for θ [ 0 , 1 ] has an upper tail dependency of λ u = θ . The corresponding centered, doubly stochastic matrix is of rank n 1 . For θ = 0 , C 0 , CA = C Π , while C 1 , CA = C M with the decompositions argued earlier. Figure 3 shows the computed singular vectors and values for θ ( 0 , 1 ) . The singular vectors in Figure 3(a) drop near u = 1 . The decay of singular values starting from s 1 is similar for the different values of θ , but it is shifted upward for higher values of θ , as shown in Figure 3(b). Figure 3(d) shows that the reconstruction is significantly improved when the MAR is used, especially for large θ . For large absolute values of θ , larger absolute values of η in the MAR are chosen (Figure 3(c)).

Figure 3 
                  Analysis of the CA checkerboard copula decompositions using the raw and MAR model for various values of 
                        
                           
                           
                              θ
                           
                           \theta 
                        
                      and 
                        
                           
                           
                              n
                              =
                              50
                           
                           n=50
                        
                     : (a) elements 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    i
                                    ,
                                    j
                                 
                              
                           
                           {u}_{i,j}
                        
                      (
                        
                           
                           
                              i
                              ∈
                              
                                 [
                                 
                                    5
                                 
                                 ]
                              
                              ,
                              j
                              ∈
                              
                                 [
                                 
                                    n
                                 
                                 ]
                              
                           
                           i\in \left[5],j\in \left[n]
                        
                     ) of the first five singular vectors for 
                        
                           
                           
                              θ
                              =
                              0.5
                           
                           \theta =0.5
                        
                     . The singular vectors have a similar course for other values of 
                        
                           
                           
                              θ
                              ∈
                              
                                 (
                                 
                                    0
                                    ,
                                    1
                                 
                                 )
                              
                           
                           \theta \in \left(0,1)
                        
                     , (b) the singular values for 
                        
                           
                           
                              θ
                              ∈
                              
                                 {
                                 
                                    0.25
                                    ,
                                    0.5
                                    ,
                                    0.75
                                 
                                 }
                              
                           
                           \theta \in \left\{0.25,0.5,0.75\right\}
                        
                     , (c) Frobenius-norm minimizing choice of 
                        
                           
                           
                              η
                           
                           \eta 
                        
                      in the MAR for approximations of rank one, and (d) Frobenius error of the MAR and raw representation for approximations of rank one.
Figure 3

Analysis of the CA checkerboard copula decompositions using the raw and MAR model for various values of θ and n = 50 : (a) elements u i , j ( i [ 5 ] , j [ n ] ) of the first five singular vectors for θ = 0.5 . The singular vectors have a similar course for other values of θ ( 0 , 1 ) , (b) the singular values for θ { 0.25 , 0.5 , 0.75 } , (c) Frobenius-norm minimizing choice of η in the MAR for approximations of rank one, and (d) Frobenius error of the MAR and raw representation for approximations of rank one.

The Gumbel family of copulas with CDF

C θ , Gu ( u , v ) = exp [ ( ln u ) θ + ( ln v ) θ ] 1 θ

for θ [ 1 , ) is an Archimedean copula and exhibits upper tail dependence like the CA copula. The checkerboard copula contains high values in the upper right part (Figures 4(a) and 5(a)). The Gumbel copula is the independence copula for θ = 1 and approaches the comonotonicity copula for θ . The singular vectors in Figure 6(a) contain jumps next to u = 1 like the singular vectors for the CA copula. Again, the approximation improves considerably when using the MAR, particularly for higher values of θ , as shown in Figure 6(c) for approximations of rank one or in Figure 6(d) for approximations of rank five. The difference in the MAR and the raw representation approximation reduces when the approximation order increases (Figure 6(b)).

Figure 4 
                  Analysis of the truncation of order 10 of a Gumbel checkerboard copula with 
                        
                           
                           
                              θ
                              =
                              2.5
                           
                           \theta =2.5
                        
                      and 
                        
                           
                           
                              n
                              =
                              50
                           
                           n=50
                        
                     : (a) the checkerboard PDF, (b) the yellow squares indicate negative matrix elements in 
                        
                           
                           
                              
                                 
                                    T
                                 
                                 
                                    10
                                 
                              
                              
                                 (
                                 
                                    
                                       
                                          A
                                       
                                       
                                          n
                                       
                                    
                                    
                                       [
                                       
                                          50
                                       
                                       ]
                                    
                                 
                                 )
                              
                           
                           {T}_{10}\left({{\bf{A}}}^{n}\left[50])
                        
                     , and (c) difference of approximation and corrected approximation, 
                        
                           
                           
                              
                                 
                                    G
                                 
                                 
                                    ‒
                                    1
                                 
                              
                              
                                 (
                                 
                                    
                                       
                                          T
                                       
                                       
                                          10
                                       
                                    
                                    
                                       (
                                       
                                          
                                             
                                                A
                                             
                                             
                                                n
                                             
                                          
                                          
                                             [
                                             
                                                50
                                             
                                             ]
                                          
                                       
                                       )
                                    
                                 
                                 )
                              
                              ‒
                              P
                              
                                 (
                                 
                                    
                                       
                                          G
                                       
                                       
                                          ‒
                                          1
                                       
                                    
                                    
                                       (
                                       
                                          
                                             
                                                T
                                             
                                             
                                                10
                                             
                                          
                                          
                                             (
                                             
                                                
                                                   
                                                      A
                                                   
                                                   
                                                      n
                                                   
                                                
                                                
                                                   [
                                                   
                                                      50
                                                   
                                                   ]
                                                
                                             
                                             )
                                          
                                       
                                       )
                                    
                                 
                                 )
                              
                           
                           {G}^{&#x2012;1}\left({T}_{10}\left({{\bf{A}}}^{n}\left[50]))&#x2012;P\left({G}^{&#x2012;1}\left({T}_{10}\left({{\bf{A}}}^{n}\left[50])))
                        
                     , using Algorithm 1. Note the different scaling compared to (a).
Figure 4

Analysis of the truncation of order 10 of a Gumbel checkerboard copula with θ = 2.5 and n = 50 : (a) the checkerboard PDF, (b) the yellow squares indicate negative matrix elements in T 10 ( A n [ 50 ] ) , and (c) difference of approximation and corrected approximation, G 1 ( T 10 ( A n [ 50 ] ) ) P ( G 1 ( T 10 ( A n [ 50 ] ) ) ) , using Algorithm 1. Note the different scaling compared to (a).

Figure 5 
                  Analysis of the truncation of order 10 of a Gumbel checkerboard copula with 
                        
                           
                           
                              θ
                              =
                              7.5
                           
                           \theta =7.5
                        
                      and 
                        
                           
                           
                              n
                              =
                              50
                           
                           n=50
                        
                     : (a) the checkerboard PDF, (b) the yellow squares indicate the negative matrix elements in 
                        
                           
                           
                              
                                 
                                    T
                                 
                                 
                                    10
                                 
                              
                              
                                 (
                                 
                                    
                                       
                                          A
                                       
                                       
                                          n
                                       
                                    
                                    
                                       [
                                       
                                          50
                                       
                                       ]
                                    
                                 
                                 )
                              
                           
                           {T}_{10}\left({{\bf{A}}}^{n}\left[50])
                        
                     . They occur more frequently than for 
                        
                           
                           
                              θ
                              =
                              2.5
                           
                           \theta =2.5
                        
                     , and (c) difference of approximation and corrected approximation, 
                        
                           
                           
                              
                                 
                                    G
                                 
                                 
                                    ‒
                                    1
                                 
                              
                              
                                 (
                                 
                                    
                                       
                                          T
                                       
                                       
                                          10
                                       
                                    
                                    
                                       (
                                       
                                          
                                             
                                                A
                                             
                                             
                                                n
                                             
                                          
                                          
                                             [
                                             
                                                50
                                             
                                             ]
                                          
                                       
                                       )
                                    
                                 
                                 )
                              
                              ‒
                              P
                              
                                 (
                                 
                                    
                                       
                                          G
                                       
                                       
                                          ‒
                                          1
                                       
                                    
                                    
                                       (
                                       
                                          
                                             
                                                T
                                             
                                             
                                                10
                                             
                                          
                                          
                                             (
                                             
                                                
                                                   
                                                      A
                                                   
                                                   
                                                      n
                                                   
                                                
                                                
                                                   [
                                                   
                                                      50
                                                   
                                                   ]
                                                
                                             
                                             )
                                          
                                       
                                       )
                                    
                                 
                                 )
                              
                           
                           {G}^{&#x2012;1}\left({T}_{10}\left({{\bf{A}}}^{n}\left[50]))&#x2012;P\left({G}^{&#x2012;1}\left({T}_{10}\left({{\bf{A}}}^{n}\left[50])))
                        
                     , using Algorithm 1. Note the different scaling compared to (a).
Figure 5

Analysis of the truncation of order 10 of a Gumbel checkerboard copula with θ = 7.5 and n = 50 : (a) the checkerboard PDF, (b) the yellow squares indicate the negative matrix elements in T 10 ( A n [ 50 ] ) . They occur more frequently than for θ = 2.5 , and (c) difference of approximation and corrected approximation, G 1 ( T 10 ( A n [ 50 ] ) ) P ( G 1 ( T 10 ( A n [ 50 ] ) ) ) , using Algorithm 1. Note the different scaling compared to (a).

Figure 6 
                  Analysis of the Gumbel checkerboard copula decompositions using the raw and MAR model for 
                        
                           
                           
                              θ
                              =
                              10
                           
                           \theta =10
                        
                      and 
                        
                           
                           
                              n
                              =
                              50
                           
                           n=50
                        
                     : (a) elements 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    i
                                    ,
                                    j
                                 
                              
                           
                           {u}_{i,j}
                        
                      (
                        
                           
                           
                              i
                              ∈
                              
                                 [
                                 
                                    5
                                 
                                 ]
                              
                              ,
                              j
                              ∈
                              
                                 [
                                 
                                    n
                                 
                                 ]
                              
                           
                           i\in \left[5],j\in \left[n]
                        
                     ) of the first singular vectors for 
                        
                           
                           
                              θ
                              =
                              10
                           
                           \theta =10
                        
                     . The continuous Gumbel copula has an upper tail dependence, (b) the Frobenius error of the approximation for a Gumbel copula with 
                        
                           
                           
                              θ
                              =
                              10
                           
                           \theta =10
                        
                      and increasing approximation order 
                        
                           
                           
                              
                                 
                                    n
                                 
                                 
                                    *
                                 
                              
                           
                           {n}^{* }
                        
                     . The MAR reduces the error considerably, (c) Frobenius error for approximations of rank one, and (d) Frobenius error for approximations of rank five.
Figure 6

Analysis of the Gumbel checkerboard copula decompositions using the raw and MAR model for θ = 10 and n = 50 : (a) elements u i , j ( i [ 5 ] , j [ n ] ) of the first singular vectors for θ = 10 . The continuous Gumbel copula has an upper tail dependence, (b) the Frobenius error of the approximation for a Gumbel copula with θ = 10 and increasing approximation order n * . The MAR reduces the error considerably, (c) Frobenius error for approximations of rank one, and (d) Frobenius error for approximations of rank five.

For higher parameter values θ , the truncated representations of the Gumbel copula contain negative entries. Figures 4 and 5 show the discretized PDF, its approximation, and an indicator plot for the invalidity of the elements. The negative elements in Figures 4(b) and 5(b) have a waveform. The Gumbel copula contains higher peaks for higher parameters θ , and the approximation tends to have more negative elements. After applying the correction algorithm, all elements are nonnegative, and the Frobenius distance between the (corrected) approximation and the discretized PDF is smaller (Table 1). The corrections are smaller for small values of θ (Figure 4(c)) than for larger values of θ (Figure 5(c)).

Table 1

Frobenius distances for the approximation of a Gumbel checkerboard copula with parameter θ and n = 50

θ = 2.5 θ = 7.5
C n [ 50 ] G 1 ( T 10 ( A n [ 50 ] ) ) F 0.0084 0.6449
C n [ 50 ] P ( G 1 ( T 10 ( A n [ 50 ] ) ) ) F 0.0084 0.5476
G 1 ( T 10 ( A n [ 50 ] ) ) P ( G 1 ( T 10 ( A n [ 50 ] ) ) ) F 0.0008 0.3099
G 1 ( T 10 ( A n [ 50 ] ) ) P ( G 1 ( T 10 ( A n [ 50 ] ) ) ) F C n [ 50 ] F 0.05% 11.47%

G 1 ( T 10 ( A n [ 50 ] ) ) denotes the truncation, and P ( G 1 ( T 10 ( A n [ 50 ] ) ) ) the result of Algorithm 1. The distance between the original and the approximation decreases with the application of Algorithm 1. The last row displays the relative change through Algorithm 1 with respect to the Frobenius norm of the raw matrix C n [ 50 ] .

3.2 Decompositions of asymmetric copulas

For asymmetric copulas, the left and right singular vectors do not coincide. We use an asymmetric construction method from Nelsen [35, p. 84], which yields copulas with cubic sections. The copula CDF is

(26) C a , b , asym ( u , v ) = u v + u v ( 1 u ) ( 1 v ) [ ( a b ) v ( 1 u ) + b ] ,

where b 1 , [ b 3 ( 9 + 6 b 3 b 2 ) 1 2 ] 2 a 1 , and a b . The conditions on a and b ensure the validity of the resulting copula. For a = b [ 1 , 1 ] the FGM copula with parameter b arises. Figure 7 shows the resulting SVD for two configurations of a and b . For a = 0.5 and b = 0.5 the resulting copula CDF is

C 0.5 , 0.5 , asym ( u , v ) = u v + u ( 1 u ) 2 v 2 ( 1 v ) 0.5 u v ( 1 u ) ( 1 v ) ,

and for a = 1.5 and b = 0.5

C 1.5 , 0.5 , asym ( u , v ) = u v 2 u v 2 ( 1 u ) 2 ( 1 v ) + 0.5 u v ( 1 u ) ( 1 v ) .

In both cases, the left and right singular vectors are the polynomials of degree two. The geometric dimension of the discretized copula is 2. Thus, the singular values in Figure 7(e) drop at 3 to zero. The singular values are larger for the first singular value combination than for the second. The left singular vectors in Figure 7(a) and (c) have similar courses but change order. The right singular vectors (Figure 7(b) and (d)) exhibit a greater variation between the combinations of parameters than the left singular values. They show y -axis mirroring but also change slope and are shifted.

Figure 7 
                  Analysis of the asymmetric checkerboard copula decomposition of the copula following equation (26) with 
                        
                           
                           
                              n
                              =
                              50
                           
                           n=50
                        
                      and two configurations of 
                        
                           
                           
                              a
                           
                           a
                        
                      and 
                        
                           
                           
                              b
                           
                           b
                        
                     . The left singular vectors are similar between the two parameter configurations, whereas the right singular values exhibit strong differences: (a) elements 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    i
                                    ,
                                    j
                                 
                              
                           
                           {u}_{i,j}
                        
                      (
                        
                           
                           
                              i
                              ∈
                              
                                 [
                                 
                                    2
                                 
                                 ]
                              
                              ,
                              j
                              ∈
                              
                                 [
                                 
                                    n
                                 
                                 ]
                              
                           
                           i\in \left[2],j\in \left[n]
                        
                     ) of the left singular vectors 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    i
                                 
                              
                           
                           {u}_{i}
                        
                      with 
                        
                           
                           
                              a
                              =
                              0.5
                           
                           a=0.5
                        
                      and 
                        
                           
                           
                              b
                              =
                              ‒
                              0.5
                           
                           b=&#x2012;0.5
                        
                     , (b) elements 
                        
                           
                           
                              
                                 
                                    v
                                 
                                 
                                    i
                                    ,
                                    j
                                 
                              
                           
                           {v}_{i,j}
                        
                      (
                        
                           
                           
                              i
                              ∈
                              
                                 [
                                 
                                    2
                                 
                                 ]
                              
                              ,
                              j
                              ∈
                              
                                 [
                                 
                                    n
                                 
                                 ]
                              
                           
                           i\in \left[2],j\in \left[n]
                        
                     ) of the right singular vectors 
                        
                           
                           
                              
                                 
                                    v
                                 
                                 
                                    i
                                 
                              
                           
                           {v}_{i}
                        
                      with 
                        
                           
                           
                              a
                              =
                              0.5
                           
                           a=0.5
                        
                      and 
                        
                           
                           
                              b
                              =
                              ‒
                              0.5
                           
                           b=&#x2012;0.5
                        
                     , (c) elements 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    i
                                    ,
                                    j
                                 
                              
                           
                           {u}_{i,j}
                        
                      (
                        
                           
                           
                              i
                              ∈
                              
                                 [
                                 
                                    2
                                 
                                 ]
                              
                              ,
                              j
                              ∈
                              
                                 [
                                 
                                    n
                                 
                                 ]
                              
                           
                           i\in \left[2],j\in \left[n]
                        
                     ) of the left singular vectors 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    i
                                 
                              
                           
                           {u}_{i}
                        
                      with 
                        
                           
                           
                              a
                              =
                              ‒
                              1.5
                           
                           a=&#x2012;1.5
                        
                      and 
                        
                           
                           
                              b
                              =
                              0.5
                           
                           b=0.5
                        
                     , (d) elements 
                        
                           
                           
                              
                                 
                                    v
                                 
                                 
                                    i
                                    ,
                                    j
                                 
                              
                           
                           {v}_{i,j}
                        
                      (
                        
                           
                           
                              i
                              ∈
                              
                                 [
                                 
                                    2
                                 
                                 ]
                              
                              ,
                              j
                              ∈
                              
                                 [
                                 
                                    n
                                 
                                 ]
                              
                           
                           i\in \left[2],j\in \left[n]
                        
                     ) of the right singular vectors 
                        
                           
                           
                              
                                 
                                    v
                                 
                                 
                                    i
                                 
                              
                           
                           {v}_{i}
                        
                      with 
                        
                           
                           
                              a
                              =
                              ‒
                              1.5
                           
                           a=&#x2012;1.5
                        
                      and 
                        
                           
                           
                              b
                              =
                              0.5
                           
                           b=0.5
                        
                     , and (e) the singular values 
                        
                           
                           
                              
                                 
                                    s
                                 
                                 
                                    i
                                 
                              
                           
                           {s}_{i}
                        
                      drop to zero after 
                        
                           
                           
                              
                                 
                                    s
                                 
                                 
                                    2
                                 
                              
                           
                           {s}_{2}
                        
                      as the geometric dimension is 2.
Figure 7

Analysis of the asymmetric checkerboard copula decomposition of the copula following equation (26) with n = 50 and two configurations of a and b . The left singular vectors are similar between the two parameter configurations, whereas the right singular values exhibit strong differences: (a) elements u i , j ( i [ 2 ] , j [ n ] ) of the left singular vectors u i with a = 0.5 and b = 0.5 , (b) elements v i , j ( i [ 2 ] , j [ n ] ) of the right singular vectors v i with a = 0.5 and b = 0.5 , (c) elements u i , j ( i [ 2 ] , j [ n ] ) of the left singular vectors u i with a = 1.5 and b = 0.5 , (d) elements v i , j ( i [ 2 ] , j [ n ] ) of the right singular vectors v i with a = 1.5 and b = 0.5 , and (e) the singular values s i drop to zero after s 2 as the geometric dimension is 2.

3.3 Gaussian copula

We end the section with the Gaussian copula and apply the notions of Section 2.6. The Gaussian copula models the dependence structure of multivariate Gaussian distributions. Let F ρ denote a bivariate Gaussian CDF with correlation ρ , variance ( 1 , 1 ) and mean ( 0 , 0 ) , the PDF by f ρ and the standard univariate Gaussian counterparts by Φ and φ , respectively. Then, the CDF of a Gaussian copula with correlation ρ [ 1 , 1 ] is given by

C ρ , G a = F ρ ( Φ 1 ( u ) , Φ 1 ( v ) ) .

Figure 8(a) and (b) shows the resulting (PDF) decompositions for the checkerboard copula. As proven in the following, the singular vectors are identical for different ρ ( 0 , 1 ) in the continuous decomposition. No noticeable differences can be observed for the singular vectors of the checkerboard approximations for different values of ρ .

Figure 8 
                  Checkerboard decomposition of the Gaussian family of copulas for 
                        
                           
                           
                              n
                              =
                              50
                           
                           n=50
                        
                     , the transformed probabilist’s Hermite polynomials, and numerical estimates for the geometric dimension: (a) elements 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    i
                                    ,
                                    j
                                 
                              
                           
                           {u}_{i,j}
                        
                      (
                        
                           
                           
                              i
                              ∈
                              
                                 [
                                 
                                    5
                                 
                                 ]
                              
                              ,
                              j
                              ∈
                              
                                 [
                                 
                                    n
                                 
                                 ]
                              
                           
                           i\in \left[5],j\in \left[n]
                        
                     ) of the singular vectors for 
                        
                           
                           
                              ρ
                              =
                              0.5
                           
                           \rho =0.5
                        
                     . No discernible difference is evident in the plots for the other 
                        
                           
                           
                              ρ
                           
                           \rho 
                        
                     , (b) the singular values 
                        
                           
                           
                              
                                 
                                    s
                                 
                                 
                                    i
                                 
                              
                           
                           {s}_{i}
                        
                      increase for larger values of 
                        
                           
                           
                              ρ
                           
                           \rho 
                        
                     , (c) the first five (normalized) transformed probabilist’s Hermite polynomials 
                        
                           
                           
                              
                                 
                                    ψ
                                 
                                 
                                    i
                                 
                              
                           
                           {\psi }_{i}
                        
                     ,  and (d) the numerical estimations of the geometric dimensions increase with the grid size and are comparable for the different values of 
                        
                           
                           
                              ρ
                           
                           \rho 
                        
                     .
Figure 8

Checkerboard decomposition of the Gaussian family of copulas for n = 50 , the transformed probabilist’s Hermite polynomials, and numerical estimates for the geometric dimension: (a) elements u i , j ( i [ 5 ] , j [ n ] ) of the singular vectors for ρ = 0.5 . No discernible difference is evident in the plots for the other ρ , (b) the singular values s i increase for larger values of ρ , (c) the first five (normalized) transformed probabilist’s Hermite polynomials ψ i , and (d) the numerical estimations of the geometric dimensions increase with the grid size and are comparable for the different values of ρ .

For a bivariate Gaussian distribution, Hill [19] showed a PDF decomposition using Hermite polynomials. The following theorem extends its results to the Gaussian copula, yielding a representation in terms of transformed Hermite polynomials. We use the representation of the probabilist’s Hermite polynomial ψ i of order i by

ψ i ( x ) = ( 1 ) i exp ( x 2 2 ) d i d x i exp ( x 2 2 ) .

Theorem 1

Let c ρ be a Gaussian copula density with parameter 1 < ρ < 1 , Φ the standard Gaussian CDF, and ψ i the probabilist’s Hermite polynomial of order i. Then,

(27) c ρ ( u , v ) = 1 + i = 1 ρ i i ! ψ i ( Φ 1 ( u ) ) ψ i ( Φ 1 ( v ) ) , u , v ( 0 , 1 ) .

Proof

c ρ ( u , v ) = f ρ ( Φ 1 ( u ) , Φ 1 ( v ) ) φ ( Φ 1 ( u ) ) φ ( Φ 1 ( v ) ) x Φ 1 ( u ) , y Φ 1 ( v ) = f ρ ( x , y ) φ ( x ) φ ( y ) Hill [26] = 1 2 π exp 1 2 ( x 2 + y 2 ) 1 + i = 1 ρ i i ! ψ i ( x ) ψ i ( y ) 1 2 π exp ( x 2 2 ) 1 2 π exp y 2 2 = i = 0 ρ i i ! ψ i ( Φ 1 ( u ) ) ψ i ( Φ 1 ( v ) ) .

Using the well-known maximal correlation property of the Gaussian distribution [25, Section 6 and references therein], the representation in equation (27) is the one obtained by canonical correlation and thus a decomposition in the sense of Section 2.6 [28].

Figure 8(c) shows the first transformed probabilist’s Hermite polynomials ψ i and Figure 8(d) the geometric dimension of Gaussian checkerboard copulas for various grid sizes n . The geometric dimension is bounded by n 1 and increases with n , whereas the continuous Gaussian copula has an infinite geometric dimension. Figure 9 shows the distance between the piecewise integrated transformed Hermite polynomials (equations (19) and (20)) and the singular vectors of the Gaussian checkerboard copula for polynomial degrees 1 to 7. The distance decreases with n for all degrees. The smaller the parameter ρ and the degree, the faster the distance decreases.

Figure 9 
                  Distance between the 
                        
                           
                           
                              d
                           
                           d
                        
                     -th piecewise integrated continuous singular vectors and 
                        
                           
                           
                              d
                           
                           d
                        
                     -th singular vector of the discretized matrix for a Gaussian copula for increasing 
                        
                           
                           
                              n
                           
                           n
                        
                      and different values of 
                        
                           
                           
                              ρ
                           
                           \rho 
                        
                     . Distance decreases for all degrees and parameters 
                        
                           
                           
                              ρ
                           
                           \rho 
                        
                      considered with 
                        
                           
                           
                              n
                           
                           n
                        
                     : (a) 
                        
                           
                           
                              ρ
                              =
                              0.25
                           
                           \rho =0.25
                        
                     , (b) 
                        
                           
                           
                              ρ
                              =
                              0.50
                           
                           \rho =0.50
                        
                     , and (c) 
                        
                           
                           
                              ρ
                              =
                              0.75
                           
                           \rho =0.75
                        
                     .
Figure 9

Distance between the d -th piecewise integrated continuous singular vectors and d -th singular vector of the discretized matrix for a Gaussian copula for increasing n and different values of ρ . Distance decreases for all degrees and parameters ρ considered with n : (a) ρ = 0.25 , (b) ρ = 0.50 , and (c) ρ = 0.75 .

3.4 Copula similarities

The difference between copulas can be quantified using the calculated measures of Section 2.5. Figures 10 and 11 show the examples of the similarity of copulas using the (normalized) Frobenius distance of discretizations of Section 2.5. Figure 10 shows the distance between Gaussian copulas with different correlations. While Figure 10(a) shows the Frobenius distance, Figure 10(b) and (c) shows the results using the normalizations. In Figure 10(b), the distances are scaled by a common factor, which results in pairs of Gaussian copulas with large ρ being considered more dissimilar than pairs with small ρ with identical difference. The second normalization, δ 2 , scales based on Pearson’s ϕ 2 of the copulas and yields similar differences for pairs of the same ρ difference. With increasing discretization grid size, n , δ 1 tends to shrink the distance, whereas δ 2 maintains the value (Figure 10(d)). Figure 11 shows the two normalizations for various checkerboard approximations of parametric copulas for two values of τ . Using the Frobenius distance or δ 1 in Figure 11(a) generally produces higher distances if at least one copula has a high τ . Normalization δ 2 produces close distance values for similar values of τ as shown in Figure 11(b) and covers a wider range of possible distances between 0 and 1. Figure 11(c) shows the results of the truncated approximation of δ 1 and δ 2 . Most of the pairs of Gaussian copulas considered already approximate the distance for small k * .

Figure 10 
                  Comparison of the normalizations of Section 2.5 for a Gaussian copula with various copula correlations 
                        
                           
                           
                              ρ
                           
                           \rho 
                        
                      and 
                        
                           
                           
                              n
                              =
                              100
                           
                           n=100
                        
                     . Normalization 
                        
                           
                           
                              
                                 
                                    δ
                                 
                                 
                                    1
                                 
                              
                           
                           {\delta }_{1}
                        
                      tends to shrink the distance with increasing 
                        
                           
                           
                              n
                           
                           n
                        
                     : (a) 
                        
                           
                           
                              
                                 
                                    
                                       ‖
                                       
                                          ⋅
                                       
                                       ‖
                                    
                                 
                                 
                                    F
                                 
                              
                           
                           {\Vert \cdot \Vert }_{F}
                        
                     , (b) 
                        
                           
                           
                              
                                 
                                    δ
                                 
                                 
                                    1
                                 
                              
                              
                                 (
                                 
                                    ⋅
                                 
                                 )
                              
                           
                           {\delta }_{1}\left(\cdot )
                        
                     . For small 
                        
                           
                           
                              ρ
                           
                           \rho 
                        
                     , the copulas are assigned similar differences, (c) 
                        
                           
                           
                              
                                 
                                    δ
                                 
                                 
                                    2
                                 
                              
                              
                                 (
                                 
                                    ⋅
                                 
                                 )
                              
                           
                           {\delta }_{2}\left(\cdot )
                        
                     , and (d) comparison the normalizations for two Gaussian copulas with 
                        
                           
                           
                              ρ
                              =
                              0.4
                           
                           \rho =0.4
                        
                      and 
                        
                           
                           
                              ρ
                              =
                              0.6
                           
                           \rho =0.6
                        
                     , respectively, with increasing grid size 
                        
                           
                           
                              n
                           
                           n
                        
                     .
Figure 10

Comparison of the normalizations of Section 2.5 for a Gaussian copula with various copula correlations ρ and n = 100 . Normalization δ 1 tends to shrink the distance with increasing n : (a) F , (b) δ 1 ( ) . For small ρ , the copulas are assigned similar differences, (c) δ 2 ( ) , and (d) comparison the normalizations for two Gaussian copulas with ρ = 0.4 and ρ = 0.6 , respectively, with increasing grid size n .

Figure 11 
                  Comparison of the normalizations of Section 2.5 for a (G)umbel, (C)layton, and (Ga)ussian copula for two different values of 
                        
                           
                           
                              τ
                           
                           \tau 
                        
                      and 
                        
                           
                           
                              n
                              =
                              100
                           
                           n=100
                        
                     . Kendall’s 
                        
                           
                           
                              τ
                           
                           \tau 
                        
                      values refer to the copula to be discretized. 11(c) uses the truncated representation of 
                        
                           
                           
                              
                                 
                                    δ
                                 
                                 
                                    1
                                 
                              
                           
                           {\delta }_{1}
                        
                      and 
                        
                           
                           
                              
                                 
                                    δ
                                 
                                 
                                    2
                                 
                              
                           
                           {\delta }_{2}
                        
                      and shows the computed values for increasing truncation parameter: (a) 
                        
                           
                           
                              
                                 
                                    δ
                                 
                                 
                                    1
                                 
                              
                              
                                 (
                                 
                                    ⋅
                                 
                                 )
                              
                           
                           {\delta }_{1}\left(\cdot )
                        
                     . For 
                        
                           
                           
                              τ
                              =
                              0.3
                           
                           \tau =0.3
                        
                     , the copulas are assigned similar differences, (b) 
                        
                           
                           
                              
                                 
                                    δ
                                 
                                 
                                    2
                                 
                              
                              
                                 (
                                 
                                    ⋅
                                 
                                 )
                              
                           
                           {\delta }_{2}\left(\cdot )
                        
                     , and (c) the truncated, normalized distances 
                        
                           
                           
                              
                                 
                                    δ
                                 
                                 
                                    1
                                 
                              
                           
                           {\delta }_{1}
                        
                      and 
                        
                           
                           
                              
                                 
                                    δ
                                 
                                 
                                    2
                                 
                              
                           
                           {\delta }_{2}
                        
                      from Section 2.5 over truncation order 
                        
                           
                           
                              
                                 
                                    k
                                 
                                 
                                    ⋆
                                 
                              
                           
                           {k}^{\star }
                        
                      for pairs of Gaussian copulas with 
                        
                           
                           
                              n
                              =
                              100
                           
                           n=100
                        
                     .
Figure 11

Comparison of the normalizations of Section 2.5 for a (G)umbel, (C)layton, and (Ga)ussian copula for two different values of τ and n = 100 . Kendall’s τ values refer to the copula to be discretized. 11(c) uses the truncated representation of δ 1 and δ 2 and shows the computed values for increasing truncation parameter: (a) δ 1 ( ) . For τ = 0.3 , the copulas are assigned similar differences, (b) δ 2 ( ) , and (c) the truncated, normalized distances δ 1 and δ 2 from Section 2.5 over truncation order k for pairs of Gaussian copulas with n = 100 .

4 Visual exploratory analysis of copulas with profile plots

A primary purpose of correspondence analysis is usually to generate visual representations of high-dimensional data by projecting row and column profiles into a low-dimensional space while maximizing the covered variation of the data (for an introduction, see, e.g., [18]). We start by describing the approach and identifying the characteristics of the copula visible in the graphs, and thus, the characteristics of the graphs to be analyzed in Section 4.1. In Section 4.2, we use empirical data plots from ranked pseudo-observations to analyze the dependence structure.

4.1 Understanding and interpreting profile plots

In profile plots, the similarity of the rows and columns of the checkerboard copula is shown. A row corresponds to the conditional distribution of u given the “row” value of v (where we use the standard notation of u being the horizontal coordinate and v being the vertical coordinate). The row profiles, F , and the column profiles, G , correspond to the singular value-weighted coordinates in the space spanned by the opposing singular vectors, i.e., in the notation of Section 2.2, F U S and G V S . All n row and column profiles are shown on the basis of their first two coordinates in the profile plot. Therefore, a row profile shows the two most significant coordinates of the rows with respect to the basis spanned by the columns and vice versa. The proximity of different row profiles reflects the similarity of the corresponding conditional distributions of u given the value of v . For example, for independent variables, the distribution of u given v does not change with the value of v , and all profile points in a profile plot would match. In a case with monotone dependence instead, the distribution of u given v changes with v , and the profile points referring to different values of v would not match, and their distance increases with the dissimilarity of the respective conditional distributions. In a 2D plot of the first two basis vectors, the v value of the row profile is not visible in the coordinates of the points. Thus, we color the profiles to reflect the position of the profile: the lighter the color, the closer the v value is to zero. Thus, the row profile for the conditional distribution u given v 0 is the point with the lightest color, and the profile given v 1 is the point with the darkest color.

The profiles of rows and columns, F ˜ and G ˜ , using the MAR can be computed analogously. We compare F and F ˜ for the Clayton copula below. For symmetric copulas, F and F ˜ differ only in singular values and not in singular vectors that lead to the same shapes but different profile scalings (Lemma 1). In traditional correspondence analysis, normalizations of the row and column profiles account for the variation of the frequencies of the individual profiles. All profiles have the same frequencies in the copula domain. Therefore, normalizations are not necessary in this setting. Although sometimes plotted in one figure, the distances between a row and a column profile cannot be interpreted directly because the representation is based on a different basis. The profile plots of rows and columns are identical if the underlying copula is symmetric positive definite; i.e., C n and ( C n ) are equal, and C n is positive definite. The matrix C n is, in particular, for exchangeable copulas, symmetric but not necessarily positive definite. A copula is exchangeable if ( F X ( X ) , F Y ( Y ) ) has the same distribution as ( F Y ( Y ) , F X ( X ) ) . The plot of several copula profiles in one plot displays differences between copulas.

Figure 12 shows the graphs for some of the copulas of Sections 3.1 and 3.2, visualizing the general characteristics depicted in the profile plot. Profiles of the raw model lying close to the zero point indicate approximately conditionally independent distributions since the most significant coordinates are close to zero. For an independence copula, all profiles lie close to zero. Significant deviations between the components in the raw model graph and the MAR model graph refer to strong characteristics of the comonotonicity copula. Figure 12(a) shows the examples for an independence and in Figure 12(b) and (c) for a comonotonicity copula. Through the points’ colors, the plots also display how the profiles evolve and how rapidly the profiles change. Points of similar colors lying close together exhibit a smooth evolution of the copula, whereas varying distances show more extensive changes of the copula in certain areas. Increasing changes are evident, for example, in the case of tail dependence, where the profiles change rapidly in the area of the tail. The plot of the comonotonicity copula in the raw model in Figure 12(b) shows unordered profiles. The comonotonicity copulas SVD is ambiguous since any orthonormal set of vectors forms singular vectors of the diagonal matrix. Thus, the calculated basis is merely random, and the profiles are scattered. For the Gumbel copula, the profiles in Figure 12(d) and (e) evolve smoothly. Still, the differences become larger for higher values of θ and the profiles closer to one since the copula has an upper tail dependence that increases with θ . Using the MAR affects the profiles only slightly in Figure 12(e) as MAR only changes singular values and not singular vectors for symmetric copulas. Figure 12(f) shows the profiles of a Gaussian copula for different values of ρ in one chart. The similarity of profiles changes most pronounced in the tails of the profiles, whereas the profile differences corresponding to middle columns and rows remain similar. Figure 12(g) depicts the row and column profiles of an asymmetric copula, where the profiles do not coincide but are mirrored with respect to the horizontal axis.

Figure 12 
                  Row and column profiles for four copulas with various parameters, each with grid size 
                        
                           
                           
                              n
                              =
                              50
                           
                           n=50
                        
                     . Except for the copula in (g), the displayed copulas are symmetric and have identical row and column profiles. The profiles reflect various copula characteristics, such as the strength of dependence, symmetry for different axes, and areas with high variation: (a) the independence copula in the raw model. All profiles lie close to zero, (b) the comonotonicity copula in the raw model. The profiles are scattered, (c) the comonotonicity copula in the MAR. All profiles lie close to zero, (d) the Gumbel copula in the raw model, (e) the Gumbel copula in the MAR. The axis limits vary slightly compared to (d) and (f) the Gaussian copula in the raw model, and (g) the asymmetric copula according to equation (26). Row and column profiles differ.
Figure 12

Row and column profiles for four copulas with various parameters, each with grid size n = 50 . Except for the copula in (g), the displayed copulas are symmetric and have identical row and column profiles. The profiles reflect various copula characteristics, such as the strength of dependence, symmetry for different axes, and areas with high variation: (a) the independence copula in the raw model. All profiles lie close to zero, (b) the comonotonicity copula in the raw model. The profiles are scattered, (c) the comonotonicity copula in the MAR. All profiles lie close to zero, (d) the Gumbel copula in the raw model, (e) the Gumbel copula in the MAR. The axis limits vary slightly compared to (d) and (f) the Gaussian copula in the raw model, and (g) the asymmetric copula according to equation (26). Row and column profiles differ.

4.2 Profile plots illustrated with a data example

Using data from an engineering context, we apply the graphical dependence assessment to empirical data. Coblenz et al. [6] modeled the distribution of fuel drops that are generated by a fuel injector in a jet engine using vine copulas. The droplets are characterized by five variables x 1 , , x 5 , i.e., the size of the drop, the position in the x and y directions, and the velocity in the x and y directions. Data are generated using numerical simulations under different operating conditions of jet engines, specified by the air velocity, the air pressure, and the thickness of the atomizing edge. Coblenz et al. [6] published statistically simulated data for different operating conditions. We focus on one of the ten operating conditions modeled, i.e., an air velocity of 90 m s 1 , an air pressure of 5 bar, and a thickness of the atomizing edge of 230 μ m since this is the largest of the provided datasets. It consists of 5,252 points in the five dimensions listed earlier.

The published data of Coblenz et al. [6] are available in the rank-transformed copula domain, which we denote by u 1 , , u 5 . Note that due to the rank transformation, all values of u j are in the discrete set { 1 5,252 , 2 2,525 , , 1 } . The copula domain’s relative frequency table, C n , is computed by counting the number of points per lattice box in I n × I n . Observations lying precisely on a grid boundary are counted for the box below. We use n = 26 as a divisor of 5,252 for the analysis so that the resulting table has 202 observations in each row and column, and dividing each cell by 202 leads to a doubly stochastic matrix. For each distinct pair of dimensions, we plot the row profiles, the column profiles, and a checkerboard copula plot of the pseudo-observations in Figure 13. We focus on five of the pairs here, and the graphs for the combinations ( u 1 , u 3 ) , ( u 1 , u 5 ) , ( u 2 , u 5 ) , ( u 3 , u 5 ) , and ( u 4 , u 5 ) are shown in Appendix D in Figure A1.

Figure 13 
                  Profile and checkerboard plots of the fuel injector spray characteristics in jet engines from Coblenz et al. [6]. The physical interpretations of the variables are drop size (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                     ), x-position (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    2
                                 
                              
                           
                           {u}_{2}
                        
                     ), y-position (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    3
                                 
                              
                           
                           {u}_{3}
                        
                     ), x-velocity (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    4
                                 
                              
                           
                           {u}_{4}
                        
                     ), and y-velocity (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     ): (a) row profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    2
                                 
                              
                           
                           {u}_{2}
                        
                      and u
                     2, (b) column profiles for variables u
                     1 and u
                     2, (c) checkerboard plot for variables u
                     1 and u
                     2, (d) row profiles for variables u
                     1 and u
                     4, (e) column profiles for variables u
                     1 and u
                     4, (f) checkerboards plot for variables u
                     1 and u
                     4, (g) row profiles for variables u
                     2 and u
                     3, (h) column profiles for variables u
                     2 and u
                     3, (i) checkerboard plot for variables u
                     2 and u
                     3, (j) row profiles for variables u
                     2 and u
                     4. A profile at (0.60, 0.27) is out of scope, (k) column profiles for variables u
                     2 and u
                     4. A profile at (0.60, 0.27) is out of scope, (l) checkerboard copula plot for variables u
                     2 and u
                     4, (m) row profiles for variables u
                     3 and u
                     4, (n) column profiles for variables u
                     3 and u
                     4, and (o) checkerboard plot for variables u
                     3 and u
                     4.
Figure 13

Profile and checkerboard plots of the fuel injector spray characteristics in jet engines from Coblenz et al. [6]. The physical interpretations of the variables are drop size ( u 1 ), x-position ( u 2 ), y-position ( u 3 ), x-velocity ( u 4 ), and y-velocity ( u 5 ): (a) row profiles for variables u 2 and u 2, (b) column profiles for variables u 1 and u 2, (c) checkerboard plot for variables u 1 and u 2, (d) row profiles for variables u 1 and u 4, (e) column profiles for variables u 1 and u 4, (f) checkerboards plot for variables u 1 and u 4, (g) row profiles for variables u 2 and u 3, (h) column profiles for variables u 2 and u 3, (i) checkerboard plot for variables u 2 and u 3, (j) row profiles for variables u 2 and u 4. A profile at (0.60, 0.27) is out of scope, (k) column profiles for variables u 2 and u 4. A profile at (0.60, 0.27) is out of scope, (l) checkerboard copula plot for variables u 2 and u 4, (m) row profiles for variables u 3 and u 4, (n) column profiles for variables u 3 and u 4, and (o) checkerboard plot for variables u 3 and u 4.

As profile points are obtained from empirical data, they deviate from their theoretical counterparts. To visualize statistical noise in the plots, we show the typical minimal and maximal values of profiles for an independence copula by a gray rectangle in the plots. The gray rectangles are obtained by sampling 5,252 realizations from an independence copula and computing their row and column profiles. The procedure is repeated 100 times. The rectangles cover 95% of the minimal and maximal point coordinates of the 100 samples in every dimension. Thus, if the profiles are outside the gray box, the underlying copula is unlikely to be the independence copula. This approach aligns with Greenacre [18], who advocates resampling methods, for example, bootstrapping, over using asymptotic results for profile values. Again, the darker the point’s color, the closer the conditional distribution’s conditioning variable is to one.

The profile plots for variables u 1 and u 2 in Figure 13(a) and (b) show that some profiles deviate from others and that there is a continuous development with the conditioning variable. For row profiles, i.e., conditional distributions u 2 given u 1 , with u 1 close to one (dark points in the row profile plot) and column profiles u 2 close to zero (light points in the column profile plot), the profiles change and indicate that the variables are not independent. The checkerboard plot in Figure 13(c) shows a peak at ( 0 , 1 ) . Further information is covered by noise. The profiles in Figure 13(d) and (e) exhibit a U-shaped pattern and are mirrored with respect to the vertical axis. Thus, the dependence has a countermonotonic characteristic that is reflected loosely in the checkerboard copula plot in Figure 13(f). While high values are apparent near ( 0 , 1 ) and ( 1 , 0 ) , the pattern in between is hard to distinguish. The row and column profiles of Figure 13(g) and (h) differ clearly. Whereas the row profiles evolve in a similar direction with stronger changes near 0, the column profiles undergo a cyclical transformation. The profiles corresponding to small and large values of u 3 are similar, and the profiles for u 3 near 0.5 are different. This pattern is a sign of U or hump-shaped dependence, which is also reflected in the checkerboard plot. For variables u 2 and u 4 , Figure 13(j) to (l) show a typical tail-dependence behavior. The profiles change rapidly for small values of u 2 and u 4 , whereas they evolve relatively smoothly for larger values. The behavior of the profile plots in Figure 13(m) and (n) is similar to variables u 2 and u 3 , but is exchanged. The row profiles undergo a cyclical transformation, while the column profiles evolve smoothly. As the U-shaped form is more apparent than for variables u 1 and u 4 , the profiles show a stronger pattern for variables u 3 and u 4 . The u-shape is distinguishable in Figure 13(o).

Overall, the row and profile plots provide at least the same amount of information as the checkerboard plots, but they are more transparent and less cluttered than the checkerboard plots.

5 Conclusion

This article analyzes truncations of SVD and correspondence analysis of checkerboard copulas. Checkerboard copulas can be mapped to doubly stochastic matrices, making it straightforward to ensure copula properties for the approximations. We find that some common copulas, for example, comonotonicity-like, have high ranks and thus are poorly represented in the straightforward SVD and that truncations can have negative elements. To account for comonotonicity-like copulas with high ranks, we adapt a representation anchored with the comonotonicity copula and show its performance in examples. We compute the nearest valid doubly stochastic matrix to correct the truncations with negative entries. We analyze the representations of statistical characteristics of copulas, such as Kendall’s τ , Spearman’s ρ , or differences between copulas through the decomposition. The truncations can be used to compute discretized versions of continuous decompositions, linking our analysis to continuous decompositions. We derive a decomposition of the Gaussian copula into transformed Hermite polynomials and show that the discretized singular vectors draw closer to the transformed Hermite polynomials with increasing grid size. We analyze correspondence analysis profile plots for copulas and show that they reveal asymmetries and non-monotonic dependence. Profile plots for various copulas are shown, and the graphical analysis is illustrated on a dataset on fuel injector spray characteristics in jet engines.

Other approaches for reducing the comonoticity-like characteristics of the copula are possible, such as using rook copulas [7] and, for empirical data, sample-dependent grid sizes [22] or anchoring with respect to other copulas while varying the sample size [13]. They need more complex fitting of the parameters and components and might use different grid sizes. Thus, we leave the comparison of these methods for further research.

In this article, we do not expand on the empirical estimation of the model. It is well known that the empirical checkerboard copula converges to the theoretical checkerboard copula. Perturbation theory analyzes the effect of noise on the results of the SVD (for a concise overview, see, e.g., [46]). The singular vectors can suffer from large fluctuations for small noise; the singular values, however, are estimated more robustly. Thus, the visual analysis in Section 4 is less prone to noise than plotting the singular vectors directly.

Although the approach can be extended to larger dimensions, it is not straightforward. The concept of the checkerboard copula is viewed in a higher dimension, for example, in Carley and Taylor [5]. There is no direct analog of SVD in three and higher dimensions, but various approaches exist (see, for example, Kolda and Bader [26] for an introduction). Copula-specific methods for modeling high-dimensional data include vine copulas [3,14,23,36] and nested Archimedean copulas [20,42], where the copulas involved could be analyzed using the methods presented here.

Acknowledgements

We thank Johan Segers for his thoughtful and constructive discussions, as well as his insightful feedback, which have greatly enhanced the quality of this article. We extend our gratitude to the two anonymous reviewers whose valuable insights significantly improved this manuscript.

  1. Funding information: We gratefully acknowledge the financial support provided by the Bischöfliche Studienförderung Cusanuswerk to JR and by the KIT publication fund for open access publishing.

  2. Author contributions: Both authors have accepted responsibility for the entire content of this manuscript and consented to its submission to the journal and have reviewed all the results and approved the final version of the manuscript. OG: conceptualization, methodology, writing, supervision. JR: conceptualization, methodology, writing, software, simulation.

  3. Conflict of interest: The authors have no conflicts of interest related to this publication.

Appendix A Calculations for the algorithms of Section 2.3

We consider the problem

P 1 ( A ) arg min B R n × n A B F 2 s.t. B 1 = 1 , B 1 = 1 ,

with a symmetric matrix A in Appendix A.1 and an asymmetric A in Appendix A.2.

The solution of P 1 ( A 1 ) has a closed form, if the matrix A 1 is symmetric, i.e.,

P 1 sym ( A 1 ) = A 1 + 1 n I n 1 n A 1 + 1 n 2 1 A 1 1 I n 1 1 1 n 1 1 A 1 .

In the case of an asymmetric matrix A 1 , the problem P 1 ( A 1 ) boils down to a linear system (Appendix A.2).

For P 2 ( A 2 ) ,

P 2 ( A 2 ) = arg min B R n × n A 2 B F 2 s.t. B i , j 0 , i , j [ n ] ,

there exists a closed-form solution independent of the symmetry of A 2 . As the Frobenius norm can be minimized elementwise,

B A 2 F 2 = i , j : ( A 2 ) i j < 0 ( B A 2 ) i j 2 + i , j : ( A 2 ) i j 0 ( B A 2 ) i j 2 B R n × n ,

the solution of P 2 ( A 2 ) is the elementwise positive part of A 2 . Algorithm 1 combines P 1 and P 2 .

A.1 Symmetric copula

The proof is analogous to Zass and Shashua [47] for symmetric A . We provide it here for completeness and to emphasize its inapplicability to asymmetric matrices. Let A = A and P 1 * ( A ) be the relaxation

P 1 * ( A ) arg min B R n × n A B F 2 s.t. B 1 1 + B 1 1 = 0 .

If B = B , B 1 1 + B 1 1 = 0 B 1 = 1 = B 1 . Thus, if B P 1 * ( X ) = B , B is also the solution of P 1 ( X ) . Let L ( A , μ ) be the Lagrangian of P 1 * ( A ) with

L ( B , μ ) = trace ( B B ) trace ( 2 A B ) 2 μ ( B 1 + B 1 21 ) .

Then,

(A1) L ( B , μ ) B = 2 B 2 A 2 μ 1 21 μ = ! 0 B = A + μ 1 + 1 μ ( ) B = A + 1 μ + μ 1

and

(A2) B + B = ( A + A ) + 2 μ 1 + 2 1 μ B 1 + B 1 = 2 1 follows from L ( B , μ ) μ = 0 = ( A + A ) 1 + 2 μ 1 1 = n + 2 1 μ 1

(A3) μ = 1 n I n 1 2 n 1 1 I n 1 2 ( A + A ) 1 ,

using ( n I n + 1 1 ) 1 = 1 n ( I n + 1 2 n 1 1 ) . Plugging μ from equation (A3) into (A1) yields

(A4) B = A + 1 n ( I n 1 2 n 1 1 ) ( I n 1 2 ( A + A ) ) 1 1 + 1 1 n ( I n 1 2 n 1 1 ) ( I n 1 2 ( A + A ) ) 1 = A + 1 n I n 1 2 n ( A + A ) + 1 2 n 2 1 ( A + A ) 1 I n 1 1 1 2 n 1 1 ( A + A ) = A = A A + 1 n I n 1 n A + 1 n 2 1 A 1 I n 1 1 1 n 1 1 A .

The result for B in equation (A4) is symmetric, and thus, also solution for P 1 ( A ) :

B = A + 1 n I n 1 n A + 1 n 2 1 A 1 I n 1 1 1 n 1 1 A = A + 1 n I n 1 n A + 1 n 2 1 A 1 I n 1 1 1 n 1 1 A = B ,

using

( A 1 1 ) i j = k = 1 n A i k = k = 1 n A k j = ( 1 1 A ) i j , i , j [ n ] .

Algorithm 1: Algorithm to compute the nearest doubly stochastic matrix in terms of the Frobenius error following Zass and Shashua [47] for symmetric matrices A and ε > 0 . The stopping criterion c c max ensures the termination of the algorithm. The solutions of P 1 and P 2 can be found in Appendix A.
input Matrix A R n × n , ε > 0 , number of maximal iterations c max
output: nearest doubly stochastic matrix B
1 Set B = A and c = 1 ;
1 Update B = P 1 ( B ) ;
3 while i , j : B i , j < ε c c max do
4 5 6 Update B = P 2 ( B ) ; Update B = P 1 ( B ) ; Update c = c + 1 ;
7 end

A.2 Asymmetric copula

For asymmetric A , the result of P 1 * ( A ) is not symmetric and thus is not a solution to the original problem P 1 ( A ) .

Instead, the solution of a Karush-Kuhn-Tucker equation system yields the solution for P 1 . The problem

P 1 ( A ) arg min B R n × n A B F 2 s.t. B 1 = 1 , B 1 = 1

with the Lagrange function and its derivative

L ( B , λ , μ ) = trace ( B B ) trace ( 2 A B ) λ ( B 1 1 ) μ ( B 1 1 ) L B i j ( B , λ , μ ) = 2 B i j 2 A i j λ j μ i , for i , j [ n ]

yields the system

(A5) 2 B i j λ j μ i = 2 A i j , ( i , j ) [ n ] × [ n ] ,

(A6) i = 1 n B i j = 1 , j [ n ] ,

(A7) j = 1 n B i j = 1 , i [ n ] .

The solution of the Karush-Kuhn-Tucker equation system is the solution of the linear equation system K b = a with

K ˜ = 2 I n [ n 2 × n 2 ] [ 1 I n [ n × n ] ] [ I n [ n × n ] 1 ] [ 1 I n [ n × n ] ] 0 0 [ I n [ n × n ] 1 ] 0 0 , b ˜ = B λ μ , a ˜ = 2 A 0 n × 1 0 n × 1

and K = K ˜ 1 : n 2 + 2 n 1 , 1 : n 2 + 2 n 1 , b = b ˜ 1 : n 2 + 2 n 1 , and a = a ˜ 1 : n 2 + 2 n 1 . Thereby, denotes the Kronecker product of the matrices, and denotes the column-wise stacking of a matrix into a vector. The last row and column are excluded, as the matrix K is singular and the constraint j = 1 n B n j = 1 is guaranteed by the remaining constraints (A6) and (A7). Then, the first n 2 elements of the solution b rearranged as matrix B are the solution of P 1 .

Additionally, Algorithm 1 includes a deflection component to account for the more general setting [17,47], as summarized in Algorithm 2.

Algorithm 2: Algorithm to compute the nearest doubly stochastic matrix in terms of the Frobenius error following Zass and Shashua [47] for asymmetric matrices A and ε > 0 . The stopping criterion c c max ensures the termination of the algorithm. The solutions of P 1 and P 2 can be found in A.2. B c , d denotes the solution in iteration c in the subproblem d and I c , d the corresponding incremental change
input: Matrix A R n × n , ε > 0 , number of maximal iterations c max
output: Nearest doubly stochastic matrix B
1 Set B = A and c = 1 ;
2 Set B 1 , 1 = P 1 ( B ) and I 1 , 1 = B 1 , 1 B ;
3 Set B 1 , 2 = P 2 ( B + I 1 , 1 ) and I 1 , 2 = B c , 2 ( B + I c , 1 ) ;
4 repeat
5 6 7 Update c = c + 1 ; Set B c , 1 = P 1 ( B + I c 1 , 2 ) and I c , 1 = B c , 1 ( B + I c 1 , 2 ) ; Set B c , 2 = P 2 ( B + I c , 1 ) and I c , 2 = B c , 2 ( B + I c , 1 ) ;
8 ( B + I c , 1 + I c , 2 ) ε ( B + I c , 1 + I c , 2 ) 1 1 < ε ( B + I c , 1 + I c , 2 ) 1 1 < ε c c max

B Decomposition in terms of the Hellinger distance

The SVD and the algorithms of Section 2.3 yield minimal errors in terms of the Frobenius norm. The SVD is also the best low-rank approximation considering the spectral norm [34]. In statistics, the Hellinger distance is often used to assess the proximity of densities (see [1,32], for two recent contributions). In this section, we analyze Hellinger distance-based decompositions for two different versions of the Hellinger distance for matrices, as, to our knowledge, there is no agreed definition in the matrix case yet. For a matrix square root-based Hellinger distance, the decomposition generalizes from the Frobenius case, while it is of a different and more complicated structure for an elementwise square root Hellinger distance.

For discrete probability distributions p = ( p 1 , , p n ) and q = ( q 1 , , q n ) , the Hellinger norm d H is computed by

d H ( p , q ) = 1 2 i = 1 n ( p i q i ) 2 1 2 = 1 2 i = 1 n ( p i + q i ) 2 i = 1 p i q i 1 2 .

For matrices, there are different notions of the Hellinger distance in the literature. We consider a formulation based on the matrix square root first [4] and then turn to an elementwise square root method [39].

Bhatia et al. [4] started from the decomposition of the Hellinger distance for densities into an arithmetic and geometric mean. As the geometric mean for matrices can be interpreted in various ways, different notions of the distance can be obtained. We use their most straightforward generalization yielding the Hellinger distance for positive semidefinite, and thus, symmetric, matrices A and B

d H ( A , B ) = A 1 2 B 1 2 F = [ trace ( A + B ) 2 trace ( A 1 2 B 1 2 ) ] 1 2 .

Thereby, A 1 2 denotes the matrix square root with ( A 1 2 ) A 1 2 = A 1 2 A 1 2 = A .

Lemma 3

The low-rank approximation problem of a positive definite matrix A yields the same eigenvectors and eigenvalues for the Frobenius and the Hellinger distance.

Proof

Let P F ( A , k ) the low-rank approximation problem in terms of the Frobenius distance

P F ( A , k ) min B R n × n A B F s.t. rank ( B ) k

and P H ( A , k ) in terms of the Hellinger distance

(A8) P H ( A , k ) = min B R n × n A 1 2 B 1 2 F s.t. rank ( B ) k , B positive semidefinite ,

as the matrix square root is unique and rank ( B ) = rank ( B 1 2 ) [21, Theorem 7.2.6].

Thus, the minimizing argument, B 1 2 , of Problem P H ( A , k ) is the k -truncated SVD of A 1 2 . Due to the positive definiteness of A , left and right singular values are identical, and the eigenvalue decomposition

(A9) A 1 2 U Λ ( U )

exists with eigenvector matrix U and diagonal eigenvalue matrix Λ . This yields for the matrix square root of the minimizing argument, B 1 2 , of P H ( A , k ) and the minimizing argument, B,

B 1 2 = U [ k ] Λ [ k ] ( U [ k ] ) and B = B 1 2 B 1 2 = U [ k ] Λ [ k ] 2 ( U [ k ] ) .

The eigenvectors of A 1 2 and A are identical and the singular vectors are squared, as

A = A 1 2 A 1 2 = ( U Λ U ) ( U Λ U ) = U Λ 2 U .

Thus, the minimizing argument of P F ( A , k ) is

U [ k ] Λ [ k ] 2 ( U [ k ] )

and equal to the minimizing argument of P H ( A , k ) .□

This definition of the Hellinger distance obtains the same decomposition as with the Frobenius distance. The coefficient η in the MAR can also be computed with the Hellinger distance instead of the Frobenius distance in equation (8). However, the definition of the Hellinger distance only for positive semidefinite copulas is restrictive, as valid bistochastic matrices do not need to be positive definite, for example, unsymmetric.

Rao [39] and Cuadras and Cuadras [11] defined the Hellinger distance in terms of an elementwise square root, thus considering only matrices with non-negative elements. Let A denote the elementwise square root of a matrix A . Then, the decomposition based on the elementwise Hellinger decomposition is for a symmetric checkerboard copula C n

(A10) C n = U H Λ H ( U H ) .

Truncations T n * ( C n ) have to be squared elementwise to obtain a low-rank approximation of the Hellinger decomposition. Note that the squared decomposition does not keep the rank of T n * ( A n ) . The MAR could be used in the elementwise Hellinger decomposition, and the optimization in equation (8) could be adapted and solved by a general optimization problem solver. To our knowledge, no optimizations similar to those in the Frobenius case are available for the elementwise Hellinger scenario, as either the objective function contains square roots or the constraints are non-linear. While the elements of the squared decomposition are non-negative, the row and column sums are not one, in general, and thus, the squared decomposition is not doubly stochastic.

All in all, the elementwise Hellinger decomposition is not as straightforward as the Frobenius decomposition, as the squared decomposition does not keep the rank of the truncation, and the attached optimization problems obtain more complex. Through the elementwise square root, the influence of peaks in the checkerboard copula on the objective function is reduced compared to the Frobenius case. It is a modeling choice, whether this is desired or not. Rao [39] and Cuadras and Cuadras [11] pointed out that elementwise Hellinger-based decomposition’s main advantage is the independence from the row and column marginals. However, the marginals are constant in the checkerboard copula setting; thus, the correspondence analysis does not depend on them. Thus, we do not expand on the Hellinger decompositions in the main part of this article.

C Computations for Spearman’s ρ and Kendall’s τ in Section 2.4

As in Section 2.4, A n = U S V denotes the centered copulas SVD. Let additionally be u 0 = v 0 = 1 n 1 and s 0 = 1 to ease the notation. The equations for Spearman’s ρ in Section 2.4 follow from

ρ ( C n ) 3 n trace ( Ω C n ) 3 = 3 n trace Ω k = 0 n 1 u k s k v k 3 = 3 n k = 0 n 1 s k trace ( Ω u k v k ) 3 = 3 ω ˇ 2 n k = 0 n 1 s k trace ( ω ω u k v k ) 3 = 3 ( 4 3 n 1 3 n ) n k = 0 n 1 s k trace ( ω u k v k ω ) 3 = 4 1 n 2 k = 0 n 1 s k ω , u k v k , ω 3 = 4 1 n 2 * ω , 1 n 1 2 + k = 1 n * s k ω , u k v k , ω 3 = 4 1 n 2 k = 1 n 1 s k ω , u k v k , ω + 3 3 = 4 1 n 2 k = 1 n 1 s k ω , u k v k , ω .

Similarly, follows for the MAR decomposition, A ˜ n = U ˜ S ˜ ( V ˜ ) ,

ρ ( C n ) = 3 n trace Ω k = 1 n 1 u ˜ k s ˜ k v ˜ k η I n + ( 1 + η ) Π n 3 = 4 1 n 2 k = 1 n 1 s ˜ k trace ( ω ω u ˜ k v ˜ k ) η trace ( ω ω I n ) + ( 1 + η ) trace ( ω ω Π n ) 3 = 4 1 n 2 k = 1 n 1 s ˜ k u ˜ k , ω v ˜ k , ω η ω 2 + ( 1 + η ) * ω , 1 n 1 2 3 = 4 1 n 2 k = 1 n 1 s ˜ k u ˜ k , ω v ˜ k , ω 4 1 n 2 η + 3 n ( 1 + η ) 3 = ( * ) 4 1 n 2 k = 1 n 1 ( s k + η ) u ˜ k , ω 2 4 1 n 2 η + 3 n ( 1 + η ) 3 ,

where ( * ) is only valid for symmetric A ˜ n .

The respective computations for Kendall’s τ are

τ ( C n ) 1 1 n 2 trace ( E C n E ( C n ) ) = 1 1 n 2 k 1 = 0 n 1 k 2 = 0 n 1 s k 1 s k 2 trace ( E u k 1 v k 1 E v k 2 u k 2 ) = 1 1 n 2 k 1 = 0 n 1 k 2 = 0 n 1 s k 1 s k 2 trace ( u k 2 E u k 1 v k 1 E v k 2 ) = 1 1 n 2 k 1 = 0 n 1 k 2 = 0 n 1 s k 1 s k 2 u k 2 , E u k 1 v k 1 , E v k 2

and for the MAR analogously.

D Further figures for Section 4.2

Figure A1

Figure A1 
                  Remaining profile and checkerboard plots of the fuel injector spray characteristics in jet engines from Coblenz et al. [6] from Section 4.2. The other dimension combinations are shown in Figure 13. The physical interpretations of the variables are drop size (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                     ), x-position (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    2
                                 
                              
                           
                           {u}_{2}
                        
                     ), y-position (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    3
                                 
                              
                           
                           {u}_{3}
                        
                     ), x-velocity (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    4
                                 
                              
                           
                           {u}_{4}
                        
                     ), and y-velocity (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     ). For the variable pairs (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                     , 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     ) and (
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    2
                                 
                              
                           
                           {u}_{2}
                        
                     , 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     ), no deviation from independence is discernible. A weak hump-shape can be observed for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    3
                                 
                              
                           
                           {u}_{3}
                        
                     . Again, the course of column profiles is reversed in the middle of the profiles. The plots show a Gaussian-like behavior for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    3
                                 
                              
                           
                           {u}_{3}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     . The profile plots for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    4
                                 
                              
                           
                           {u}_{4}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                      show a weak deviation from independence for the profiles near 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    4
                                 
                              
                              =
                              1
                           
                           {u}_{4}=1
                        
                      and extreme values of 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     : (a) row profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    3
                                 
                              
                           
                           {u}_{3}
                        
                     , (b) column profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    3
                                 
                              
                           
                           {u}_{3}
                        
                     , (c) checkerboard plot for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    3
                                 
                              
                           
                           {u}_{3}
                        
                     , (d) row profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , (e) column profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , (f) checkerboard plot for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    1
                                 
                              
                           
                           {u}_{1}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , (g) row profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    2
                                 
                              
                           
                           {u}_{2}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , (h) column profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    2
                                 
                              
                           
                           {u}_{2}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , (i) checkerboard plot for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    2
                                 
                              
                           
                           {u}_{2}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , (j) row profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    3
                                 
                              
                           
                           {u}_{3}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , (k) column profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    3
                                 
                              
                           
                           {u}_{3}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , (l) checkerboard plot for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    3
                                 
                              
                           
                           {u}_{3}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , (m) Row profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    4
                                 
                              
                           
                           {u}_{4}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , (n) column profiles for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    4
                                 
                              
                           
                           {u}_{4}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     , and (o) checkerboard plot for variables 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    4
                                 
                              
                           
                           {u}_{4}
                        
                      and 
                        
                           
                           
                              
                                 
                                    u
                                 
                                 
                                    5
                                 
                              
                           
                           {u}_{5}
                        
                     .
Figure A1

Remaining profile and checkerboard plots of the fuel injector spray characteristics in jet engines from Coblenz et al. [6] from Section 4.2. The other dimension combinations are shown in Figure 13. The physical interpretations of the variables are drop size ( u 1 ), x-position ( u 2 ), y-position ( u 3 ), x-velocity ( u 4 ), and y-velocity ( u 5 ). For the variable pairs ( u 1 , u 5 ) and ( u 2 , u 5 ), no deviation from independence is discernible. A weak hump-shape can be observed for variables u 1 and u 3 . Again, the course of column profiles is reversed in the middle of the profiles. The plots show a Gaussian-like behavior for variables u 3 and u 5 . The profile plots for variables u 4 and u 5 show a weak deviation from independence for the profiles near u 4 = 1 and extreme values of u 5 : (a) row profiles for variables u 1 and u 3 , (b) column profiles for variables u 1 and u 3 , (c) checkerboard plot for variables u 1 and u 3 , (d) row profiles for variables u 1 and u 5 , (e) column profiles for variables u 1 and u 5 , (f) checkerboard plot for variables u 1 and u 5 , (g) row profiles for variables u 2 and u 5 , (h) column profiles for variables u 2 and u 5 , (i) checkerboard plot for variables u 2 and u 5 , (j) row profiles for variables u 3 and u 5 , (k) column profiles for variables u 3 and u 5 , (l) checkerboard plot for variables u 3 and u 5 , (m) Row profiles for variables u 4 and u 5 , (n) column profiles for variables u 4 and u 5 , and (o) checkerboard plot for variables u 4 and u 5 .

References

[1] Aya-Moreno, C., Geenens, G., Penev, S. (2018). Shape-preserving wavelet-based multivariate density estimation. Journal of Multivariate Analysis, 168, 30–47, DOI: https://doi.org/10.1016/j.jmva.2018.07.002. 10.1016/j.jmva.2018.07.002Search in Google Scholar

[2] Bakam, Y. I. N., Pommeret, D. (2022). K-Sample test for equality of Copulas. Search in Google Scholar

[3] Bedford, T., Cooke, R. M. (2002). Vines-a new graphical model for dependent random variables. The Annals of Statistics, 30(4), 1031–1068, DOI: https://doi.org/10.1214/aos/1031689016. 10.1214/aos/1031689016Search in Google Scholar

[4] Bhatia, R., Gaubert, S., Jain, T. (2019). Matrix versions of the Hellinger distance. Letters in Mathematical Physics, 109(8), 1777–1804. 10.1007/s11005-019-01156-0Search in Google Scholar

[5] Carley, H., Taylor, M. D. (2002). A new proof of Sklar’s theorem. In: Distributions with given marginals and statistical modelling (pp. 29–34), Netherlands: Springer. 10.1007/978-94-017-0061-0_4Search in Google Scholar

[6] Coblenz, M., Holz, S., Bauer, H.-J., Grothe, O., Koch, R. (2020). Modelling fuel injector spray characteristics in jet engines by using vine copulas. Journal of the Royal Statistical Society Series C: Applied Statistics, 69(4), 863–886. 10.1111/rssc.12421Search in Google Scholar

[7] Cottin, C., Pfeifer, D. (2014). From Bernstein polynomials to Bernstein copulas. Journal of Applied Functional Analysis, 9, 277–288. Search in Google Scholar

[8] Cuadras, C. M. (2015). Contributions to the diagonal expansion of a bivariate copula with continuous extensions. Journal of Multivariate Analysis, 139, 28–44. 10.1016/j.jmva.2015.02.015Search in Google Scholar

[9] Cuadras, C. M. (2002). Correspondence analysis and diagonal expansions in terms of distribution functions. Journal of Statistical Planning and Inference, 103(1–2), 137–150. 10.1016/S0378-3758(01)00216-6Search in Google Scholar

[10] Cuadras, C. M., Augé, J. (1981). A continuous general multivariate distribution and its properties. Communications in Statistics - Theory and Methods, 10(4), 339–353. 10.1080/03610928108828042Search in Google Scholar

[11] Cuadras, C. M., Cuadras, D. (2006). A parametric approach to correspondence analysis. Linear Algebra and its Applications, 417(1), 64–74. 10.1016/j.laa.2005.10.029Search in Google Scholar

[12] Cuadras, C. M., Díaz, W. (2012). Another generalization of the bivariate FGM distribution with two-dimensional extensions. Acta et Commentationes Universitatis Tartuensis de Mathematica 16(1), 3–12. 10.12697/ACUTM.2012.16.01Search in Google Scholar

[13] Cuberos, A., Masiello, E., Maume-Deschamps, V. (2020). Copulas checker-type approximations: Application to quantiles estimation of sums of dependent random variables. Communications in Statistics - Theory and Methods, 49(12), 3044–3062. 10.1080/03610926.2019.1586936Search in Google Scholar

[14] Czado, C. (2019). Analyzing Dependent Data with Vine Copulas, vol. 222. Cham: Springer International Publishing. 10.1007/978-3-030-13785-4Search in Google Scholar

[15] Durante, F., Sempi, C. (2015). Principles of Copula Theory. New York: Chapman and Hall/CRC.10.1201/b18674Search in Google Scholar

[16] Durrleman, V., Nikeghbali, A., Roncalli, T. (2000). Copulas Approximation and New Families. DOI: https://doi.org/10.2139/ssrn.1032547.10.2139/ssrn.1032547Search in Google Scholar

[17] Dykstra, R. L. (1983). An algorithm for restricted least squares regression. Journal of the American Statistical Association, 78(384), 837–842. 10.1080/01621459.1983.10477029Search in Google Scholar

[18] Greenacre, M. J. (1984). Theory and applications of correspondence analysis. London: Academic Press. Search in Google Scholar

[19] Hill, M. O. (1974). Correspondence analysis: A neglected multivariate method. Applied Statistics, 23(3), 340–354. 10.2307/2347127Search in Google Scholar

[20] Hofert, M., Hofert, M., Mächler, M. (2011). Nested Archimedean copulas meet R: The nacopula package. Journal of Statistical Software, 39(9), 1–20. 10.18637/jss.v039.i09Search in Google Scholar

[21] Horn, R. A., Johnson, C. R. (2012). Matrix Analysis. 2nd edition, Cambridge; New York: Cambridge University Press. 10.1017/CBO9781139020411Search in Google Scholar

[22] Janssen, P., Swanepoel, J., Veraverbeke, N. (2012). Large sample behavior of the Bernstein copula estimator. Journal of Statistical Planning and Inference, 142(5), 1189–1197, DOI: https://doi.org/10.1016/j.jspi.2011.11.020. 10.1016/j.jspi.2011.11.020Search in Google Scholar

[23] Joe, H. (1996). Families of m-variate distributions with given margins and m(m‒1)⁄2 bivariate dependence parameters. Lecture Notes-Monograph Series, 28, 120–141. 10.1214/lnms/1215452614Search in Google Scholar

[24] Kazmierczak, J.-B. (1978). Migrations interurbaines dans la banlieue sud de paris. Cahiers de laanalyse des données, 3(2), 203–218. Search in Google Scholar

[25] Klaassen, C. A. J., Wellner, J. A. (1997). Efficient estimation in the bivariate normal copula model: normal margins are least favourable. Bernoulli, 3(1), 55, DOI: https://doi.org/10.2307/3318652. 10.2307/3318652Search in Google Scholar

[26] Kolda, T. G., Bader, B. W. (2009). Tensor decompositions and applications. SIAM Review, 51(3), 455–500. 10.1137/07070111XSearch in Google Scholar

[27] Kolesárová, A., Mesiar, R., Mordelová, J., Sempi, C. (2006). Discrete copulas. IEEE Transactions on Fuzzy Systems, 14(5), 698–705. 10.1109/TFUZZ.2006.880003Search in Google Scholar

[28] Lancaster, H. O. (1957). Some properties of the bivariate normal distribution considered in the form of a contingency table. Biometrika, 44(1–2), 289–292, DOI: https://doi.org/10.1093/biomet/44.1-2.289.10.1093/biomet/44.1-2.289Search in Google Scholar

[29] Li, X., Mikusiński, P., Sherwood, H., Taylor, M. D. (1997). On approximation of copulas. In Beneš, V., Štěpán, J., editors, Distributions with given Marginals and Moment Problems (pp. 107–116). Netherlands: Springer. 10.1007/978-94-011-5532-8_13Search in Google Scholar

[30] Masuhr, A., Trede, M. (2020). Bayesian estimation of generalized partition of unity copulas. Dependence Modeling, 8(1), 119–131, DOI: https://doi.org/10.1515/demo-2020-0007. 10.1515/demo-2020-0007Search in Google Scholar

[31] Mayor, G., Suner, J., Torrens, J. (2005). Copula-like operations on finite settings. IEEE Transactions on Fuzzy Systems, 13(4), 468–477, DOI: https://doi.org/10.1109/TFUZZ.2004.840129. 10.1109/TFUZZ.2004.840129Search in Google Scholar

[32] Meier, C., Kirch, C., Meyer, R. (2018). Bayesian nonparametric analysis of multivariate time series: a matrix gamma process approach. Journal of Multivariate Analysis, 175, 104560, DOI: https://doi.org/10.1016/j.jmva.2019.104560. 10.1016/j.jmva.2019.104560Search in Google Scholar

[33] Mesiar, R., Najjari, V. (2014). New families of symmetric/asymmetric copulas. Fuzzy Sets and Systems, 252, 99–110. 10.1016/j.fss.2013.12.015Search in Google Scholar

[34] Mirsky, L. (1960). Symmetric Gauge functions and unitarily invariant norms. The Quarterly Journal of Mathematics, 11(1), 50–59. 10.1093/qmath/11.1.50Search in Google Scholar

[35] Nelsen, R. B. (2006). An Introduction to Copulas. Springer Series in Statistics. New York, NY: Springer New York. Search in Google Scholar

[36] Panagiotelis, A., Czado, C., Joe, H., Stöber, J. (2017). Model selection for discrete regular vine copulas. Computational Statistics & Data Analysis, 106, 138–152. 10.1016/j.csda.2016.09.007Search in Google Scholar

[37] Perfect, H., Mirsky, L. (1965). Spectral properties of doubly-stochastic matrices. Monatshefte für Mathematik, 69(1), 35–57, DOI: https://doi.org/10.1007/BF01313442. 10.1007/BF01313442Search in Google Scholar

[38] Pfeifer, D., Tsatedem, H. A., Mändle, A., Girschig, C. (2016). New copulas based on general partitions-of-unity and their applications to risk management. Dependence Modeling, 4(1), 000010151520160006, DOI: https://doi.org/10.1515/demo2016-0006. 10.1515/demo-2016-0006Search in Google Scholar

[39] Rao, C. R. (1995). A review of canonical coordinates and an alternative to correspondence analysis using Hellinger distance. Qüestiió, 19(1-2-3), 23–63. Search in Google Scholar

[40] Rodríguez-Lallena, J. (2004). A new class of bivariate copulas. Statistics & Probability Letters, 66(3), 315–325. 10.1016/j.spl.2003.09.010Search in Google Scholar

[41] Rontsis, N., Goulart, P. (2020). Optimal approximation of doubly stochastic matrices. In: Chiappa, S., Calandra, R., editors, Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, volume 108 of Proceedings of Machine Learning Research (pp. 3589–3598). Search in Google Scholar

[42] Savu, C., Trede, M. (2010). Hierarchies of Archimedean copulas. Quantitative Finance, 10(3), 295–304. 10.1080/14697680902821733Search in Google Scholar

[43] Savu, C., Trede, M. (2008). Goodness-of-fit tests for parametric families of Archimedean copulas. Quantitative Finance, 8(2), 109–116, DOI: https://doi.org/10.1080/14697680701207639. 10.1080/14697680701207639Search in Google Scholar

[44] Schmid, F., Schmidt, R., Blumentritt, T., Gaißer, S., Ruppert, M. (2010). Copula-based measures of multivariate association. In Jaworski, P., Durante, F., Härdle, W. K., Rychlik, T., editors, Copula Theory and Its Applications (vol. 198, pp. 209–236). Springer Berlin Heidelberg. 10.1007/978-3-642-12465-5_10Search in Google Scholar

[45] Sklar, A. (1959). Fonctions de répartition à n dimensions et leurs marges. Publications de L’Institut de Statistique de L’Université de Paris, 8, 229–231. Search in Google Scholar

[46] Stewart, G. W. (1991). Perturbation theory for the singular value decomposition. In: Vaccaro, R. J., editor, and University of Rhode Island, SVD and Signal Processing, II: Algorithms, Analysis, and Applications, Amsterdam; New York: New York, N.Y., U.S.A: Elsevier; Distributors for the U.S.A. and Canada, Elsevier Science Pub. Co., (pp. 99–109). Search in Google Scholar

[47] Zass, R., Shashua, A. (2007). Doubly stochastic normalization for spectral clustering. In: Schölkopf, B., Platt, J., Hofmann, T., editors, Advances in Neural Information Processing Systems (vol. 19, pp. 1569–1576). The MIT Press. 10.7551/mitpress/7503.003.0201Search in Google Scholar

Received: 2023-11-17
Revised: 2024-05-03
Accepted: 2024-07-08
Published Online: 2024-09-18

© 2024 the author(s), published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 29.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/demo-2024-0006/html
Scroll to top button