Home Mathematics On the Bivariate Generalized Gamma-Lindley Distribution
Article Open Access

On the Bivariate Generalized Gamma-Lindley Distribution

  • Mohamed Ali Ghorbel EMAIL logo and Rahmouna Mecene
Published/Copyright: March 31, 2023
Become an author with De Gruyter Brill

ABSTRACT

In this research paper, we propose a new class of bivariate distributions called Bivariate Generalized Gamma-Lindley (BGGL) distribution with four parameters. This model is a mixture of independent Gamma random variables and bivariate generalized Lindley distribution. We investigate various properties of the new bivariate distribution such as graphical representation, joint moments and correlation. Furthermore, we derive a measure of entropy of this bivariate distribution. We also derive the distributions of the random variables X1 + X2, X1/(X1 + X2), X1/X2 and X1X2 as well as the corresponding moment properties when X1 and X2 follow the BGGL distribution. Additionally, we address two approximations of the product of the proposed model and assess their goodness of fit. Next, we elaborate the expectation maximization (E.M) algorithm in order to estimate the BGGL model parameters. Finally, we provide two concrete examples to demonstrate the applicability of the results.

2020 Mathematics Subject Classification: Primary 60E05; 62E15; 62E17

1. Introduction

Statistical distributions are intrinsic in terms of describing and predicting real world phenomena. They are frequently used in life data analysis and reability engineering. However, many of these classical distributions do not provide enough flexibility for analysing different types of lifetime data and are not sufficient for modelization in several other areas such the discipline of signal processing applications or the area of image processing and computer vision (for more details, see [15]). Therefore, a new class of distribution was introduced by Lindley [12] to analyse failure time data, especially in application modeling stress-strength reability (see, for example, [5]). This new class of distribution can be also used in a wide variety of fields including biology, engineering and medicine. The one-parameter Lindley distribution is defined by its probability density function (pdf) expressed as

1 f(x;θ)=θ2(1+x)eθx1+θ,x>0,θ>0.

It has been explored and generalized by numerous authors. Basically, Ghitany et al. [5] provided a treatment of the mathematical properties of equation (1). Next, Shanker et al. [22] derived a two-parameter Lindley distribution of which the one parameter Lindley distribution is a particular case, and they demonstrated its applicability in modeling waiting and survival times data. In [26], Zakerzadeh and Dolati introduced a three-parameter generalization of the one-parameter Lindley distribution and examined multiple properties of this new distribution. Furthermore, Torabi et al. [24] set forward a four parameters extension of the generalized Lindley distribution. In addition, discrete versions of the Lindley distribution were proposed in literature by Gómez-Déniz and Calderín-Ojeda [7], Gómez-Déniz et al. [8], Hussain et al. [10] and many other authors. Zeghdoudi and Nedjar [27] proposed another extension of Lindley distribution called Gamma Lindley distribution which offers a more flexible model for modeling lifetime data and they investigated its properties. Recently, Laribi et al. [13] derived a new generalized Gamma-Lindley distribution through mixing the Gamma and the three-parameters generalized Lindley distributions. Its pdf is (for x, α, θ > 0, γ ≥ 0 and β ≥ 1) indicated by

2 f(x;α,β,θ,γ)=θα+1xα1eθxβΓ(α+1)(γ+θ)(α+(βγ+βθθ)x),

where Γ(α):=0xα1exdx is the Euler Gamma function.

In this paper, we generate a new class of bivariate distributions called Bivariate Generalized Gamma-Lindley (BGGL) distribution. The joint pdf of this new distribution corresponds to

3 fΘ(x1,x2)=θ2α+1(x1x2)α1eθ(x1+x2)β(γ+θ)Γ2(α+1)(α2+θ(βγ+βθθ)x1x2),

where x1, x2, α, θ > 0, γ ≥ 0, β ≥ 1 and Θ := (α, β, γ, θ).

The paper is organized as follows: In Section 2, we display some properties of our distribution such as joint moments, correlation and mixture construction of the BGGL distribution defined by equation (3). Additionally, for different values of parameters, we provide graphically a variety of forms of the BGGL density. In Section 3, we derive the exact densities and moment functions of the random variables Y = X1 + X2, Z = X1/(x1 + x2), V = X1/X2 and W = X1X2, where (X1, X2) follows the BGGL distribution. Furthermore, we exhibit approximations for the distribution of W = X1X2, and discuss evidence for their robustness. The objective consists in providing simple approximations in terms of the beta and inverted beta distributions so as to facilitate the use of the known procedures for inference, prediction, etc. In Section 4, we determine the exact forms of the Rényi and Shannon entropies for the BGGL distribution. In Section 5, we investigate the estimation of the BGGL distribution parameters by the E.M algorithm. In the last section, we provide an application of the results within the field of precipitation data about rain and snow from France and we tackle another application in the field of demographic studies in USA.

2. Properties

In this section, we handle several outstanding properties of the BGGL distribution. This bivariate distribution is based essentially on a mixture of the product of two independent Gamma random variables (denoted by gamma(α, θ)) and bivariate generalized Lindley distribution (denoted by BGL(α,γ, θ)), introduced by Zakerzadeh and Dolati [26]. Both distributions are defined by the joint pdfs

fgg(x1,x2;α,θ)=θ2α(x1x2)α1Γ2(α)eθ(x1+x2),x1,x2,α,θ>0,

and

fBL(x1,x2;α,θ,γ)=θ2α+1(x1x2)α1(α2+γθx1x2)(γ+θ)Γ2(α+1)eθ(x1+x2),

where x1, x2, α, θ > 0, γ ≥ 0. Let (Y1, Y2) and (Z1, Z2) be two random vectors, such that Y1 and Y2 are independent random variables distributed according to gamma(α + 1, θ) and (Z1, Z2) is a vector distributed according to BGL(α, γ, θ). For β ≥ 1, consider the random pair (X1, X2) which stands for the mixture between (Y1, Y2) with probability (β1)β, and (Z1, Z2) with probability 1β. This mixture law is called BGGL distribution and the joint pdf of (X1, X2) is obtained by using the total probability law:

4 fΘ(x1,x2)=(β1)βfgg(x1,x2;α+1,θ)+1βfBL(x1,x2;α,θ,γ)=θ2α+1(x1x2)α1eθ(x1+x2)β(γ+θ)Γ2(α+1)(α2+θ(βγ+βθθ)x1x2),

where x1, x2, α, θ > 0, γ ≥ 0 and Θ := (α, β, γ, θ), which coincides with equation (3). Note that, the joint pdf (3) may be written in terms of the gamma function as:

fΘ(x1,x2)=θ(α2+x1x2(γθ+(β1)(γ+θ)θ))β(γ+θ)α2fg(x1;α,θ)fg(x2;α,θ).

It is easy to demonstrate that the univariate marginal density functions of equation (3) are of the form (2). When β = 1, the BGGL distribution reduces to the bivariate generalized Lindley distributions with three parameters (see [26]). For β = 1 and α = 1, the BGGL distribution reduces to the bivariate generalized Lindley distributions with two parameters (see [23]). For β = 1 and γ = 0, the random variables X1 and X2 become independent and the joint pdf (3) reduces to the product of two gamma density functions with the same parameters.

In addition, Figure 1 depicts a few graphs of the pdf defined by equation (3) for different values of parameters. Here, one can record the wide range of forms that result from the BGGL distribution.

Figure 1. Plots of the pdf of BGGL distribution for some selected values of parameters (α, β, γ, θ) as: (1) (0.05, 1.2, 1.5, 0.3); (2) (0.05, 1.2, 1.5, 2) ; (3) (2, 1.2, 1.5, 0.3); (4) (2, 1.2, 1.5, 2) ; (5) (7, 1.2, 1.5, 0.3); (6) (7, 1.2, 1.5, 3).
Figure 1.

Plots of the pdf of BGGL distribution for some selected values of parameters (α, β, γ, θ) as: (1) (0.05, 1.2, 1.5, 0.3); (2) (0.05, 1.2, 1.5, 2) ; (3) (2, 1.2, 1.5, 0.3); (4) (2, 1.2, 1.5, 2) ; (5) (7, 1.2, 1.5, 0.3); (6) (7, 1.2, 1.5, 3).

Furthermore, simple calculations yield that for each positive integer m and n, the joint (m, n)-th moment of the variables X1 and X2 having BGGL distribution, can be stated as

5 E(X1mX2n)=Γ(α+m)Γ(α+n)β(γ+θ)Γ2(α+1)θm+n((βθ+βγθ)(α+n)(α+m)+θα2).

Now, substituting appropriately for m and n in equation (5), we obtain

E(X1X2)=((βθ+βγθ)(α+1)2+θα2)β(γ+θ)θ2.

Besides, we can easily show that the correlation coefficient of X1 and X2 is indicated by

corr(X1,X2)=θ(βθ+βγθ)(β21)θ2+αβ2(γ+θ)2+β2γ(γ+2θ).

It is straightforward to notice that if α → 0, β = 1 and θ → ∞ or if α → 0, γ = 0 and β = 1, we have Corr(X1, X2) → 1/2; which corresponds also to the maximum value of the correlation for this family. The BGGL distribution stands for a mixture of i.i.d random variables since the BGL distribution is itself a mixture of independent random variables. Relying upon the result of Drouet-Mari and Kotz [3: Section 3.2.8], we infer that X1 and X2 are positively correlated by mixtures.

A bivariate distribution is said to be positively likelihood ratio dependent if the density fΘ(x1, x2) satisfies

fΘ(x1,y1)fΘ(x2,y2)fΘ(x1,y2)fΘ(x2,y1)

for all x1 > x2 and y1 > y2 (see Lehmann [14]). For the equation (3), the above inequality reduces to

(α2+θ(βγ+βθθ)x1y1)(α2+θ(βγ+βθθ)x2y2)(α2+θ(βγ+βθθ)x1y2)(α2+θ(βγ+βθθ)x2y1),

or x2y2 + x1y1x1y2 + x2y1, which clearly holds. Hence, the BGGL distribution is positively likelihood ratio dependent.

3. Some transformations

In this section, we derive explicit expressions for the pdfs and moments of Y = X1 + X2, Z = X1/(X1 + X2), V = X1/X2 and W = X1X2 when the random vector (X1, X2) is set according to BGGL distribution. In fact, for given random variables X1 and X2, the distributions of the product, sum, and ratio arise in many fields as biology, economics, engineering, genetics, medicine, number theory, physics, etc. (see for example [17]). For instance, if X1 and X2 denote the rainfall intensity and the duration of a storm, then W = X1X2 will represent the amount of rainfall produced by that storm. Another expressive example is the following: if X1 and X2 denote the proportion of days with rain and the proportion of days with snow, then Y = X1 + X2 will represent the proportion of days with precipitation (rain or snow). Moreover, here is an example of the quotient V = X1/X2 which consists in representing X1 as the crude divorce rate and X2 as the crude marriage rate. Then, V will represent the divorce to marriage ratio. Furthermore, we set forward approximations for the distribution of W = X1X2, and discuss evidence for their robustness.

3.1. PDFS

Theorems 14 and Corollary 2 yield the pdfs of Y, Z, V and W when X1 and X2 are distributed according to equation (3).

Theorem 1.

If X1 and X2 are jointly distributed according to equation (3), then the pdf of Y is expressed by

6 fY(y)=θ2α+1y2α1eθyβ(γ+θ)Γ(2α)(1+(βθ+βγθ)θy22α(2α+1)),y>0,

and the pdf of Z is given by

7 fZ(z)=zα1(1z)α1Γ(2α)β(γ+θ)Γ2(α+1)(2α(2α+1)(βθ+βγθ)z(1z)+α2θ),z(0,1).

Proof. Substituting x1 = yz and x2 = y(1 – z) with the Jacobian J(x1, x2y, z) = y, in the joint pdf of X1 and X2, we obtain the joint pdf of Y and Z as

8 fY,Z(y,z)=θ2α+1zα1(1z)α1y2α1eθyβ(γ+θ)Γ2(α+1)[ (βθ+βγθ)θy2z(1z)+α2 ],

where z ∈ (0,1) and y > 0. Now, integrating this equation appropriately, we obtain marginal densities of Y and Z.

Remark 1.

Simple calculations reveal that equation (8) can be rewritten as

fY,Z(y,z)=θβ(γ+θ)fg(y;2α,θ)fb(z;α,α)+(βθ+βγθ)β(γ+θ)fg(y;2α+2,θ)fb(z;α+1,α+1),

where fg (y; a, b) and fb (z; a, b) denote the pdfs of the gamma and the beta distributions with the parameters a and b, respectively, for all 0 < z < 1 and y > 0.

Corollary 2.

If X1 and X2 are jointly distributed according to equation (3), then the pdf of the variable V = X1/X2 is expressed by

9 fV(v)=vα1Γ(2α)β(γ+θ)(1+v)2αΓ2(α+1)(2α(2α+1)(βθ+βγθ)v(1+v)2+α2θ),v>0.

Proof. We can state that

X1X1+X2=X1/X2X1/X2+1

which, in terms of Z = X1/(X1 + X2) and V = X1/X2, can be written as Z = V/(1 + V). Therefore, making the transformation Z = V/(1 + V) with the Jacobian J(zv) = (1 + v)-2 in the pdf of Z given in (7), we get the pdf of V.

In the following, we derive the exact pdf of the product W = X1X2. The calculations involve the modified Bessel function of the second kind defined by:

Kv(x)=π(Iv(x)Iv(x))2sin(vπ),

where Iν(·) denotes the modified Bessel function of the first kind of order ν defined by

Iv(x)=xv2vΓ(v+1)k=01(v+1)kk!(x24)k.

Furthermore, K0(·) is interpreted as the limit K0(x)=limν0Kν(x). We also need the following important lemma:

Lemma 3

([18: Equation 2.3.16.1]). For p > 0 and q > 0,

0xα1epxq/xdx=2(qp)α/2Kα(2pq).
Theorem 4.

If X1 and X2 are jointly distributed according to equation (3), then the pdf of the variable W = X1X2 is provided by

10 fW(w)=2θ2α+1wα1((βθ+βγθ)θw+α2)β(γ+θ)Γ2(α+1)K0(2θw),w>0.

Proof. From (3), the joint pdf of (X1, W) = (X1, X1X2) with the Jacobian J(x1,x2x1,w)=x11, becomes

fX1,W(x1,w)=θ2α+1wα1((βθ+βγθ)θw+α2)β(γ+θ)Γ2(α+1)x11eθ(x1+wx1).

Hence, integrating this equation with respect to x1 we find that

11 fW(w)=θ2α+1wα1((βθ+βγθ)θw+α2)β(γ+θ)Γ2(α+1)0x11eθ(x1+wx1)dx1.

Direct application of Lemma 3 implies that

12 0x11eθ(x1+wx1)dx1=2K0(2θw).

The result of the theorem follows by combining equations (11) and (12). □

By using the mathematical software MATLAB, we can present numerically in Figure 2, the graphs of pdfs of Y and V defined by (6) and (9), respectively, for different values of parameters. The graph of W will be given later.

Figure 2. PDF of Y for the choices of α, β, θ and fixed parameter γ = 2 (a) and PDF of V for the choices of α, β and fixed parameters θ = 4, γ = 3 (b).
Figure 2.

PDF of Y for the choices of α, β, θ and fixed parameter γ = 2 (a) and PDF of V for the choices of α, β and fixed parameters θ = 4, γ = 3 (b).

3.2. Moments

In this part, we derive the moments of the variables Y = X1 + X2, Z = X1/(X1 + X2), V = X1/X2 and W = X1X2 defined in the previous section. We establish the following:

Theorem 5.

If X1 and X2 are jointly distributed according to equation (3), then the moment functions of the variables Y, Z, V and W are, respectively,

13 E(Yn)=k=0n(kn)Γ(nk+α)Γ(α+k)β(γ+θ)Γ2(α+1)θn((βθ+βγθ)(α+nk)(α+k)+θα2),

14 E(Zn)=αΓ(2α)Γ(α+n)β(γ+θ)Γ(α+1)Γ(2α+n)(2(βθ+βγθ)(2α+1)(α+n)(2α+n)(2α+n+1)+θ),

15 E(Vn)=Γ(α+n)Γ(αn)β(γ+θ)Γ2(α+1)((βθ+βγθ)(α2n2)+θα2),α>n,

16 E(Wn)=θ2nΓ2(α+n)β(γ+θ)Γ2(α+1)((α+n)2(βθ+βγθ)+θα2).

Proof. Firstly, we have

17 E[ (X1+X2)n ]=k=0n(kn)E(X1nkX2k).

By using (5), it is clear that

18 E(X1nkX2k)=Γ(nk+α)Γ(α+k)β(γ+θ)Γ2(α+1)θn((βθ+βγθ)(α+nk)(α+k)+θα2).

Then, substituting (18) into (17), the result in (13) follows. Next, using (7), we obtain

E(Zn)=Γ(2α)β(γ+θ)Γ2(α+1)01zn+α1(1z)α1(2α(2α+1)(βθ+βγθ)z(1z)+θα2)dz.

After performing simple calculations, we obtain (14). Similarly, equation (15) follows by using the pdf of V given in (9) and the fact that

0xa1(1+x)abdx=Γ(a)Γ(b)Γ(a+b)forx>0anda,b>0.

Finally, the result in equation (16) follows from (5) by taking n = m. □

In order to measure the symmetry and tailedness of V, W, Y and Z, we compute their skewness and kurtosis. Indeed, for the different selected values of α, β and fixed values γ = 3 and θ = 5, the distributions of V and W are highly skewed (skewness > 1), the kurtosis is greater than 3 and the distributions are leptokurtic. However, the distribution of Y is approximately symmetric (the skewness is between −1/2 and 1/2) and platykurtic since the kurtosis is less than 3. The distribution of Z is symmetric for the three cases and approximately mesokurtic (kurtosis ≈ 3) for greater values of α. The numerical values of the skewness and the kurtosis of V, W, Y and Z with their variance and ordinary moments (μr, r = 3, 4) are given in Table 1.

Table 1.

Numerical values for the ordinary moments, variance, skewness and kurtosis of V, W, Y and Z for some selected parameter values.

Moments of V, W, Y and Z for α = 5 and β = 1.5.
Transformations V W Y Z
μ3 2.2500 0.8909 1.7312
μ4 23.7163 3.3723 8.2386 0.0011
Variance 0.7762 0.6672 2.7464 0.0207
Skewness 3.2900 1.6348 0.3804
Kurtosis 39.3608 7.5758 1.0923 2.5871
Moments of V, W, Y and Z for α = 5 and β = 3.
Transformations V W Y Z
μ3 1.8748 0.9590 1.8009
μ4 16.9959 3.7467 9.0536 0.0010
Variance 0.7182 0.7164 2.9349 0.0200
Skewness 3.0801 1.5815 0.3582
Kurtosis 32.9477 7.3002 1.0511 2.5965
Moments of V, W, Y and Z for α = 10 and β = 4.
Transformations V W Y Z
μ3 0.2156 10.2281 5.9593 0.1110 × 10-5
μ4 0.5741 98.1441 74.2882 0.0003
Variance 0.2624 4.3751 9.8565 0.0110
Skewness 1.6040 1.1177 0.1926 0.9582 × 10-13
Kurtosis 8.3383 5.1274 0.7647 2.7594

3.3. Approximations of the product

In this section, we derive the approximate distribution of the product W = X1X2. It is clear that the random variable P := 1/(1 + W) has support in the interval (0, 1). For this reason, in what follows we shall account for the approximate distribution by the beta distribution with parameters a, b > 0 (say P~dβ(a,b)) and pdf

19 fP(y)=Γ(a+b)Γ(a)Γ(b)ya1(1y)b1.

The idea of approximating distributions including complicated formulae with the beta distribution is frequently tackled in the statistics literature, based on the interesting work of Dasgupta [1] (see also [4,11,16] and [21]). Note that recently, Ghorbel [6] developed this single beta distribution of the form (19) for the bivariate beta distribution. Subsequently, this transformation was opted for owing to its popularity, simplicity and its association with the derived distribution. The basic reason for doing this doesn’t lie in the fact that the calculation of equation (10) cannot be handled; it rather resides in providing a simple approximation in terms of the beta distribution so that one can use the known procedures for inference, prediction, etc. Furthermore, the presented approximation seems to be useful especially to practitioners who not only avoid the use of the modified Bessel function of the second kind but also regard the beta distribution as widely accessible in standard statistical packages.

The choice of the beta parameters a and b is enacted using the method of moments. The first two moments of P can be written as

E(P)=aa+bandE(P2)=a(a+1)(a+b)(a+b+1).

Solving simultaneously the two expressions above, it is straightforward to find that

20 a=E(P)E(P)E(P2)E(P2)E2(P)

and

21 b=(1E(P))E(P)E(P2)E(P2)E2(P).

where the two moments E(P) and E(P2) can be computed numerically by evaluating the two integrals

I1=0fW(w)1+wdwandI2=0fW(w)(1+w)2dw

for the given parameters α, β, γ and θ, where fW (w) is presented by (10). To demonstrate the closeness of the proposed approximation, we use equations (20) and (21) with eight selected values of the parameters (α, β, γ, θ). Both integrals as well as the corresponding estimated values of the parameters a and b are depicted in Table 2.

Table 2.

Estimates of (I1, I2, a, b) for selected (α, β, γ, θ).

α β γ θ I1 I2 a b
0.5 1.5 0.5 2.2 0.8662 0.7807 2.4314 0.3757
1.5 1.5 1.5 6 0.9087 0.8339 8.2760 0.8318
4 1.5 4 1.5 0.1269 0.0234 1.8010 12.3888
5 1.5 3 3 0.2700 0.0886 3.1271 8.4546
4 1.5 0.5 3.2 0.4106 0.1960 3.2063 4.6032
2 1.5 2 2 0.4666 0.2650 1.9894 2.2745
3.5 1.5 0.5 7 0.7849 0.6309 8.1810 2.2418
1 1.5 2 2 0.6835 0.5200 2.1181 0.9808

The corresponding graphics are outlined in Figure 3, demonstrating comparison between exact and approximate densities of P = 1/(1 + W). The exact pdf corresponds to the solid curve and approximate pdf stands for the broken curve. It is evident that the approximate density is quite close to the exact density. Note that the exact pdf of P is stated by (with 0 < p < 1)

fP(P)=2θ2α+1(1p)α1(θ(βθ+βγθ)(1p)p1+α2)β(γ+θ)Γ2(α+1)pα+1K0(2θ1pp),

where K0(·) is interpreted as the limit (as ν → 0) of the modified Bessel function of the second kind Kν(x). Similar findings were recorded when this exercise was repeated for many other combinations of (α, β, γ, θ). Finally, It is worth noting that the numerical results are obtained by the use of besselk function in mathematical software MATLAB.

Figure 3. The exact pdf of P (solid curve) and its approximated function (broken curve) for β = 1.5 and (1): (α, γ, θ) = (0.5, 0.5, 2.2) ; (2): (α, γ, θ) = (1.5, 1.5, 6) ; (3): (α, γ, θ) = (4, 4, 1.5) ; (4): (α, γ, θ) = (5, 3, 3); (5): (α, γ, θ) = (4, 0.5, 3.2); (6): (α, γ, θ) = (2, 2, 2); (7): (α, γ, θ) = (3.5, 0.5, 7); (8): (α, γ, θ) = (1, 2, 2).
Figure 3.

The exact pdf of P (solid curve) and its approximated function (broken curve) for β = 1.5 and (1): (α, γ, θ) = (0.5, 0.5, 2.2) ; (2): (α, γ, θ) = (1.5, 1.5, 6) ; (3): (α, γ, θ) = (4, 4, 1.5) ; (4): (α, γ, θ) = (5, 3, 3); (5): (α, γ, θ) = (4, 0.5, 3.2); (6): (α, γ, θ) = (2, 2, 2); (7): (α, γ, θ) = (3.5, 0.5, 7); (8): (α, γ, θ) = (1, 2, 2).

Afterwards, based on the performed relationship in the first approximation, we present another direct approximation of W by the inverted beta distribution of the two parameters p > 0 and q > 0 say β′ (p, q), defined by the pdf

22 Γ(a+b)Γ(a)Γ(b)xp1(1+x)(p+q),x>0.

From this, we use the method of moments to determine the parameters p and q. The first two moments of W become

E(W)=qp1andE(W2)=q(q+1)(p1)(p2).

After some algebraic manipulation, it is straightforward to verify that

23 p=1+E(W2)E(W)E(W2)E2(W)

and

24 q=E(W)E(W2)E(W)E(W2)E2(W),

where both moments E(W) and E(W2) follow from (16) by taking n = 1 and n = 2, respectively.

In order to corroborate the robustness as well as goodness of fit of the second approximation, we use (23) and (24) with given selected values of the parameters (α, β, γ, θ). The corresponding estimated values of the parameters p and q are summarized in Table 3. Afterwards, the notion of robustness is assessed by comparing the exact and the approximated pdfs of W as exhibited in (10) and (22). These comparisons are illustrated in Figure 4.

Table 3.

Estimates of (a, b) for selected (α, β, γ, θ)

α β γ θ p q
1 1.5 0.5 1.7 1.9208 0.7820
0.5 1.5 0.5 2 1.3898 0.1153
2 1.5 0.5 2.5 1.9962 0.9918
0.5 1.5 1 1 2.1309 1.7906
2.5 1.5 2.5 2 2.7739 4.6442
4 1.5 1.5 4 2.4129 1.8223
2 1.5 0.5 2 2.3193 2.0888
1.5 1.5 0.7 2 2.0428 1.1144
Figure 4. The exact pdf of W (solid curve) and its approximated function (broken curve) for β = 1.5 and (1): (α, γ, θ) = (1, 0.5, 1.7); (2): (α, γ, θ) = (0.5, 0.5, 2); (3): (α, γ, θ) = (2, 0.5, 2.5) ; (4): (α, γ, θ) = (0.5, 1, 1) ; (5): (α, γ, θ) = (2.5, 2.5, 2) ; (6): (α, γ, θ) = (4, 1.5, 4); (7): (α, γ, θ) = (2, 0.5, 2) ; (8): (α, γ, θ) = (1.5, 0.7, 2).
Figure 4.

The exact pdf of W (solid curve) and its approximated function (broken curve) for β = 1.5 and (1): (α, γ, θ) = (1, 0.5, 1.7); (2): (α, γ, θ) = (0.5, 0.5, 2); (3): (α, γ, θ) = (2, 0.5, 2.5) ; (4): (α, γ, θ) = (0.5, 1, 1) ; (5): (α, γ, θ) = (2.5, 2.5, 2) ; (6): (α, γ, θ) = (4, 1.5, 4); (7): (α, γ, θ) = (2, 0.5, 2) ; (8): (α, γ, θ) = (1.5, 0.7, 2).

4. Entropies

The entropy is a crucial concept in such areas such as classical thermodynamics, where it was first recognized, statistical mechanics and physics, information theory and many other fields. It stands for an excellent tool to quantify the amount of uncertainty involved in the value of a random variable or the outcome of a random process. Several measures of entropy have been examined and compared in literature. The simplest known entropy is the Shannon entropy (see [20]). Let (X,,P) be a probability space. Consider a pdf f associated with P, dominated by σ-finite measure μ on X. Thus, the Shannon entropy is defined by

HSH(f)=Xf(x)logf(x)dμ.

One of the main extensions of the Shannon entropy was established by Rényi (see [19]). This generalized entropy measure is expressed in terms of

25 HR(μ,f)=logG(μ)1μ(forμ>0andμ1),

where

G(μ)=Xfμdμ.

Note that the Shannon entropy is the particular case of the Rényi entropy for μ ↑ 1. The calculations of these entropies involve the hypergeometric function identified as

3F0(a,b,c;;x):=k=0(a)k(b)k(c)kxkk!,|x|<1,

where (i)k := i(i + 1) … (i + k – 1) denotes the ascending factorial.

Before deriving these entropies, we need the following lemma:

Lemma 6.

Let g(α,β,γ,θ)=limμ1h(μ), where

h(μ)=ddμ3F0(μ,μ(α1)+1,μ(α1)+1;;βθ+βγθθα2μ2).

Therefore,

26 g(α,β,γ,θ)=(2α)(βθ+βγθ)θα.

Proof. Expanding 3F0 in a series form, h(μ) is equal to

27 ddμk=0(1)kΔk(μ)k!(βθ+βγθθα2)k=k=0(1)k(βθ+βγθ)kk!(θα2)k[ ddμΔk(μ) ],

where

Δk(μ)=(1)kΓ(μ+k)Γ2(μ(α1)+1+k)Γ(μ)Γ2(μ(α1)+1)μ2k=Γ(1+μ)Γ2(μ(α1)+1+k)Γ(1+μk)Γ2(μ(α1)+1)μ2k.

Now, differentiating the logarithm of Δk (μ) with respect to μ, we get

28 ddμΔk(μ)=Δk(μ)[Ψ(1+μ)+2(α1)(Ψ(μ(α1)+1+k))Ψ(μ(α1)+1)Ψ(1+μk)2kμ],

where Ψ(·) = Γ′(·) /Γ(·) is the digamma function. Finally, substituting equation (28) into equation (27) and taking μ → 1, we get

g(α,β,γ,θ)=k=01(βθ+βγθ)kk!(θα2)k(1)kΓ2(α+k)Γ(2k)Γ2(α)×[ Ψ(2)+2(α1)(Ψ(α+k)Ψ(α))Ψ(2k)2k ].

From this equation, the result follows. □

Theorem 7.

Consider the random pair (X1, X2) defined by its pdf (3). The Rényi and the Shannon entropies are indicated by

HR(μ,f)=11μ [ (3μ2)log(θ)+2logΓ(μ(α1)+1)μlog(β) 2(μ(α1)+1)log(μ)μlog(γ+θ)2μlogΓ(α)+log3F0(μ,μ(α1)+1,μ(α1)+1;;βθ+βγθθα2μ2) ]

and

HSH(f)=3logθ2(α1)Ψ(α)+logβ+2α+log(γ+θ)+2logΓ(α)θθ(2β)βγg(α,β,γ,θ),

respectively, where g (α, β, γ, θ) is provided by equation (26).

Proof. For μ > 0 and μ ≠ 1, using the joint pdf of X1 and X2 presented by (3), we have

G(μ)=00fΘμ(x1,x2)dx1dx2=α2μθμ(2α+1)βμ(γ+θ)μΓ2μ(α+1)0x1μ(α1)eθμx1×0x2μ(α1)eθμx2(1+(βθ+βγθ)θx1α2x2)μdx2dx1.

Hence, using [9: Equation (3.383.5)], we obtain

G(μ)=α2μθμ(2α+1)βμ(γ+θ)μΓ2μ(α+1)k=0(μ(α1)+1)k(μ)kk!(α2θμ)k((βθ+βγθ)θ)k×Γ(μ(α1)+1)(θμ)μ(α1)10x1μ(α1)+keθμx1dx1.

Therefore,

G(μ)=θ3μ2Γ2(μ(α1)+1)βμ(γ+θ)μμ2μ(α1)+2Γ2μ(α)×k=0(μ)k(μ(α1)+1)k(μ(α1)+1)kk!(βθ+βγθθα2μ2)k.

Thus,

G(μ)=θ3μ2Γ2(μ(α1)+1)βμ(γ+θ)μμ2μ(α1)+2Γ2μ(α)×3F0(μ,μ(α1)+1,μ(α1)+1;;βθ+βγθθα2μ2).

Considering the logarithm of G(μ) and using (25), we get the Rényi entropy. The Shannon entropy is obtained from Rényi entropy by taking μ ↑ 1 and using L’Hopital’s rule.

5. Estimation by Expectation-Maximization (E.M) Algorithm

The Expectation-Maximization (E.M) algorithm (see [2]) is a prominent way of finding maximum-likelihood estimates (MLEs) for model parameters when data are incomplete. It is an itetarive method to approximate the maximum likelihood function. Each iteration of the E.M algorithm involves two steps. In the first one (E-Step), we compute the conditional expectation denoted by (Θ|Θ(j)), where Θ(j) = (α(j), β(j), γ(j), θ(j)) is the estimated parameters vector of Θ = (α, β, γ, θ) at the jth iteration. In the second step of the E.M alghorithm (M-Step), we maximize (Θ|Θ(j)) with respect to the vector of parameters Θ.

In this section, we discuss the problem of computing MLEs of the unknown parameters of the BGGL distribution for a given random sample of size n, of the form Λ = {(x11, x12),..., (xn1, xn2 )}· Using the mixture density representation of the BGGL distribution given by (4), the incomplete likelihood function can be written as

L(Θ)=i=1n(β1βfgg(xi1,xi2,Θ1)+1βfBL(xi1,xi2,Θ2)),

where Θ1 = (α, θ) and Θ2 = (α, θ, γ).

To complete the data of this function, we associate a discrete random vector Δ, where Δ follows a bivariate Bernoulli distribution with parameters p=β1β and 1p=1β. This implies that

P(Δ=δi)=pδi(1p)1δi,

where δi ∈ {0, 1}2. The complete samples are as follows:

{ (x11,x12,δ1),,(xn1,xn2,δn) }.

Resting on the complete observations, the complete likelihood function is

Lc(Θ)=i=1n(β1βfgg(xi1,xi2,Θ1))δi(1βfBL(xi1,xi2,Θ2))1δi.

Thus, the log-likelihood function becomes

lc(Θ):=logLc(Θ)=i=1nδi(log(β1β)+log(fgg(xi1,xi2,Θ1)))+i=1n(1δi)(log(fBL(xi1,xi2,Θ2))+log(1β)).

Now, we compute the conditional expectation (Θ|Θ(j)) of the complete log likelihood function, resting upon the observed data and the current value Θ(j) = (α(j), β(j), γ(j), θ(j)) at the jth iteration. We obtain

(Θ|Θ(j))=E(lc( Θ|Λ,Θ(j))=i=1nE(δi|Λ,Θ(j))(log(β1β)+log(fgg(xi1,xi2,Θ1)))+i=1nE((1δi)|Λ,Θ(j))(log(1β)+log(fBL(xi1,xi2,Θ2))),

where

E(δi|Λ,Θ(j)):=τi(j)=(β(j)1)β(j)fgg(xi1,xi2,Θ1(j))(β(j)1)β(j)fgg(xi1,xi2,Θ1(j))+1β(j)fBL(xi1,xi2,Θ2(j)),

and

E((1δi)|Λ,Θ(j)):=ςi(j)=1τi(j).

Note that τi(j) is the posterior probability computed in the jth iteration.

In the maximization step of the E.M algorithm, we attempt to find the maximum Θ(j+1) of the function (Θ|Θ(j)) with respect to Θ=(α, β, γ, θ).

Θ(j+1)=argmaxΘ(Θ|Θ(j)),

where

29 (Θ|Θ(j))=i=1nτi(j)( log(β1β)+(2α+2)log(θ)+α(log(xi1)+log(xi2))    θ(xi1+xi2) 2log(Γ(α+1)) )+ςi(j)( log(β)+(2α+1)log(θ)    +(α1)(log(xi1)+log(xi2))+log(α2+γθxi1xi2)θ(xi1+xi2)     log(γ+θ)2log(Γ(α+1)) ).

According to Jensen’s inequality, as (Θ|Θ(j)) is increasing, the log-likelihood function is also increasing (see [2]). Differentiating equation (29) with respect to each parameter, we get the following:

α=2nlog(θ)2nΨ(α+1)+i=1n(log(xi1)+log(xi2))+2αi=1nςi(j)α2+γθxi1xi2,β=1β(β1)i=1nτi(j)1βi=1nςi(j),θ=2n(α+1)θi=1n(xi1+xi2)+i=1nςi(j)(γxi1xi2α2+γθxi1xi21γ+θ1θ),

and

γ=i=1nςi(j)(θxi1xi2α2+γθxi1xi21γ+θ),

where Ψ(.) is the digamma function. Next, we put the previous differential equations equal to zero and we use the mathematical software MATLAB to settle this system of equations numerically. From this perspective, we repeat this step until ||Θ(j+1) – Θ(j)|| > ε, where ε is a fixed threshold of convergence. Dempster et al. [2] and Wu [25] investigated the convergence properties of the E.M algorithm. Consequently, the E.M algorithm converges within a finite iterations number and gives the parameters maximum likelihood estimates. These solutions yield the E.M estimators denoted α^,β^,γ^ and θ^. Furthermore, we elaborate the following algorithm to compute the MLEs of the unknown parameters.

Algorithm

  • Step 1: Choose some initial guess values of Θ, say Θ(0) = (α(0), β(0), γ(0), θ(0)).

  • Step 2: Obtain

    Θ(1)=argmaxΘ(Θ|Θ(0)).
  • Step 3: Continue the process until convergence takes place.

6. Real data applications

We initially consider the data of rain and snow collect from five French cities and present an application of the model given by (3).

Precipitation is any type of water that forms in the Earth’s atmosphere and then drops onto the surface of the Earth. Clouds eventually get too full of water vapor, and the precipitation turns into a liquid (rain) or a solid (snow). The rain stands for precipitation that occurs in different sizes, from big, heavy drops to light ones, but snow is one of the solid types of precipitation. It is made up of water that has been frozen.

We consider data available at the website https://en.tutiempo.net/climate collected from the following five French cities (in different parts of the country) regarding the rain and snow: Dijon, Lille, Rennes, Paris and Toulouse. The dataset comprises the number of days in each year where rain and snow appeared during the period from 1990 to 2018. We consider the following variables:

  • X1: proportion of days with rain;

  • X2 : proportion of days with snow;

  • Y = X1 + X2 : proportion of days with precipitation (rain or snow).

Table 4 displays the estimates of α^,β^,γ^ and θ^, which were obtained using the E.M algorithm considered in the previous section. Furthermore, we exhibit in the same table the estimated values of the moments E(Y) for five cities.

Table 4.

Estimated values of α^,β^,γ^,θ^ and of E(Y).

City α^ β^ γ^ θ^ E(Y)
Dijon 0.7191 1.1963 5.6880 4.9055 0.5431
Lille 0.5399 1.1710 5.6946 4.1811 0.5637
Rennes 0.3396 1.0625 8.1807 3.8836 0.5338
Paris 0.4957 1.0001 7.7070 4.1536 0.5516
Toulouse 0.3318 1.1444 8.6421 4.6344 0.4431

It was reported that the proportion of days with precipitation phenomenon (rain or snow) was similar for the five cities. Toulouse city presented the lowest proportion.

The second application rests upon demographic studies. We will use the quotient as a demographic ratio of two real data, which are the crude divorce rate and the crude marriage rate. In fact, a measure of divorces in a certain population is the divorce to marriage ratio, corresponding to the ratio of the divorce rate to the marriage rate. A measure of 0.5 means, for example, that there has been one divorce for every two new marriages recorded in a given year and in a given region. We now consider the data on crude divorce and marriage rates for women and men in 2009 for 51 states in the USA. The data were published by U.S. Census Bureau. Additionally, the web

Table 5.

Estimated values of α^,β^,γ^,θ^ and of E(V).

Sex α^ β^ γ^ θ^ E(V)
Women 7.5147 1.1990 0.3970 0.5534 1.1430
Men 5.9935 1.1984 0.3177 0.4353 1.1830
site of the data is https://www2.census.gov/library/publications/2011/compendia/statab/131ed/tables/vitstat.pdf. Rates are recorded per 1, 000 women (respectively men) residing in some area. Let X1 be the crude divorce rate, namely divorce rate per 1, 000 women (respectively men), and let X2 be the crude marriage rate, i.e., marriage rate per 1, 000 women (respectively men). Hence, V = X1/X2 is the divorce to marriage ratio. Table 5 displays the estimates of α^,β^,γ^ and θ^, which were obtained using the E.M algorithm again. It was inferred from Table 5 that the divorce to marriage ratio for women is less than that for men. Furthermore, the average of divorce for a new marriage for women amounts to 1.1430 and that for a new marriage for men equals 1.1830.

Finally, we consider the experimental results using the real dataset of the rain and snow for Dijon city in the reliabilty for different statistical models such as product of two independent Gamma random variables (PG), BGL and BGGL distributions. In order to compare these models, we use the following four criteria: Akaike Information Criterion (AIC), Bayesian Information Criterion (BIC), Corrected Akaike Information Criterion (CAIC), Hannan-Quinn information criterion (HQIC) that are defined as follows:

AIC = 2log L^+2k, BIC=2logL^+klogn,CAIC=AIC+2k(k+1)nk+1,HQIC=2logL^+2klog(logn),

where k is the number of free parameters in the model, n is the sample size and L^ is the loglikelihood. For fitting a data set, the best model is a model with the smallest value of AIC, BIC, CAIC and HQIC.

Table 6 lists the Maximum Likelihood Estimates (MLEs) of the parameters from the fitted models and the values of the following statistics: AIC, BIC, CAIC and HQIC. Based on the values of these statistics, we conclude that the BGGL distribution can provide good fits for realdata. Note that similar results and interpretations can be treated for the other datasets.

Table 6.

MLEs, AICs, BICs, CAICs and HQICs for the real data set of the rain and snow for the city of Dijon.

Distr. α^ β^ γ^ θ^ AIC BIC CAIC HQIC
BGGL 0.733 1.255 5.248 4.999 -59.526 -54.197 -57.926 -57.897
PG 0.964 3.550 -30.063 -27.399 -29.619 -29.249
BGL 0.719 7.767 4.905 -10.355 -6.358 -9.432 -9.133

(Communicated by Gejza Wimmer)


Acknowledgement

The authors are grateful to the reviewers for their valuable and insightful comments that contribute to improve the quality of the manuscript.

[1] Dasgupta, P.: Two approximations for the distribution of double non-central beta, Sankhyā 30 (1968), 83–88.Search in Google Scholar

[2] Dempster, A. P.—Laird, N. M.—Rubin, D. B.: Maximum likelihood from incomplete data via the EM algorithm, J. R. Stat. Soc. Ser. B. Stat. Methodol. 39(1) (1977), 1–38.10.1111/j.2517-6161.1977.tb01600.xSearch in Google Scholar

[3] Drouet-Mari, D.—Kotz, S.: Correlation and Dependence, Imperial College Press, London, 2001.10.1142/p226Search in Google Scholar

[4] Fan, D.-Y.: The distribution of the product of independent beta variables, Comm. Statist. Theory Methods 20 (1991), 4043–4052.10.1080/03610929108830755Search in Google Scholar

[5] Ghitany, M. E.—Atieh, B.—Nadarajah, S.: Lindley distribution and its application, Math. Comput. Simul. 78(4) (2008), 493–506.10.1016/j.matcom.2007.06.007Search in Google Scholar

[6] Ghorbel, M.: On the product of the bivariate beta components, Jpn. J. Stat. Data. Sci. 2(1) (2019), 71–80.10.1007/s42081-018-0027-1Search in Google Scholar

[7] Gómez-Déniz, E.—Calderín-Ojeda, E.: The discrete Lindley distribution: properties and applications, J. Stat. Comput. Simul. 81(11) (2011), 1405–1416.10.1080/00949655.2010.487825Search in Google Scholar

[8] Gómez-Déniz, E.—Sarabia, J. M.—Balakrishnan, N.: A multivariate discrete Poisson-Lindley distribution: extensions and actuarial applications, Astin Bull. 42(2) (2012), 655–678.Search in Google Scholar

[9] Gradshteyn, I. S.—Ryzhik, I. M.: Table of Integrals, Series, and Products, 6th edition, Academic Press, New York, 2000.Search in Google Scholar

[10] Hussain, T.—Aslam, M.—Ahmad, M.: A two parameter discrete Lindley distribution, Rev. Colomb. Estad. 39(1) (2016), 45–61.Search in Google Scholar

[11] Johannesson, B.—Giri, N.: On approximations involving the beta distribution, Comm. Statist. Simulation Comput. 24(2) (1995), 489–503.10.1080/03610919508813253Search in Google Scholar

[12] Lindley, D. V.: Fiducial distributions and Bayes theorem, J. Roy. Statist. Soc. Ser. B 20 (1958), 102–107.10.1111/j.2517-6161.1958.tb00278.xSearch in Google Scholar

[13] Laribi, D.—Masmoudi, A.—Boutouria, I.: Characterization of generalized Gamma-Lindley distribution using truncated moments of order statistics, Math. Slovaca 2 (2021), 455–474.10.1515/ms-2017-0481Search in Google Scholar

[14] Lehmann, E. L.: Some concepts of dependence, Ann. Math. Statist. 37(5) (1966), 1137–1153.10.1214/aoms/1177699260Search in Google Scholar

[15] Ma, Z.—Leijon, A.: Bayesian estimation of beta mixture models with variational inference, IEEE Trans. Pattern Anal. Mach. Intell. 33(11) (2011), 2160–2173.10.1109/TPAMI.2011.63Search in Google Scholar PubMed

[16] Nadarajah, S.—Kotz, S.: Exact and approximate distributions for the product of Dirichlet components, Kybernetika 40(6) (2004), 735–744.Search in Google Scholar

[17] Nadarajah, S.: Sums, products, and ratios for the bivariate Gumbel distribution, Math. Comput. Model. 42 (2005), 499–518.10.1016/j.mcm.2005.02.003Search in Google Scholar

[18] Prudnikov, A. P.—Brychkov, Y. A.—Marichev, O. I.: Integrals and Series, vol. 1, Gordon and Breach Science Publishers, Amsterdam, 1986.Search in Google Scholar

[19] Rényi, A.: On measures of entropy and information, Proceed. 4th Berkeley Sympos. Math. Statist. Prob., University of California Press, Berkeley, California, 1961, 547–56.Search in Google Scholar

[20] Shannon, C. E.: A mathematical theory of communication, Bell Sys. Techn. J. 27 (1948), 379–423, 623–656.10.1002/j.1538-7305.1948.tb00917.xSearch in Google Scholar

[21] Sculli, D.—Wong, K. L.: The maximum and sum of two beta variables and the analysis of PERT networks, Omega 13(3) (1985), 233–240.10.1016/0305-0483(85)90061-1Search in Google Scholar

[22] Shanker, R.—Sharma, S.—Shanker, R.: A two-parameter Lindley distribution for modeling waiting and survival times data, App. Math. 4(2) (2013), 363–368.10.4236/am.2013.42056Search in Google Scholar

[23] Tassaddaq, H.—Muhammad, A.—Munir, A.: A two parameter discrete Lindley distribution, Rev. Colomb. Estad. 39(1) (2016), 45–61.10.15446/rce.v39n1.55138Search in Google Scholar

[24] Torabi, H.—Falahati-Naeini, M.—Montazeri, N. H.: An extended generalized Lindley distribution and its applications to lifetime data, J. Statist. Res. Iran 11(2) (2014), 203–222.10.18869/acadpub.jsri.11.2.203Search in Google Scholar

[25] Wu, C. J.: On the convergence properties of the EM algorithm, Ann. Stat. 11(1) (1983), 95–103.10.1214/aos/1176346060Search in Google Scholar

[26] Zakerzadeh, H.—Dolati, A.: Generalized Lindley distribution, J. Math. Ext. 3(2) (2009), 13–25.Search in Google Scholar

[27] Zeghdoudi, H.—Nedjar, S.: Gamma Lindley distribution and its application, J. Appl. Probab. Stat. 11(1) (2016), 129–138.10.16929/as/2016.923.83Search in Google Scholar

Received: 2021-08-24
Accepted: 2022-05-04
Published Online: 2023-03-31
Published in Print: 2023-04-01

© 2023 Mathematical Institute Slovak Academy of Sciences

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 15.12.2025 from https://www.degruyterbrill.com/document/doi/10.1515/ms-2023-0038/html
Scroll to top button