Home A multi-power and multi-splitting inner-outer iteration for PageRank computation
Article Open Access

A multi-power and multi-splitting inner-outer iteration for PageRank computation

  • Bing-Yuan Pu EMAIL logo , Chun Wen and Qian-Ying Hu
Published/Copyright: December 31, 2020

Abstract

As an effective and possible method for computing PageRank problem, the inner-outer (IO) iteration has attracted wide interest in the past few years since it was first proposed by Gleich et al. (2010). In this paper, we present a variant of the IO iteration, which is based on multi-step power and multi-step splitting and is denoted by MPMIO. The description and convergence are discussed in detail. Numerical examples are given to illustrate the effectiveness of the proposed method.

MSC 2010: 65C40; 65F10

1 Introduction

With the fast development of the Internet, web search engines have become one of the most important Internet tools for information retrieval. How to list the relevant web pages plays a significant role in this filed. Among thousands of search engines, Google is one of the most popular and successful. And this is mainly attributed to its effective algorithm, PageRank.

PageRank, a link-based algorithm formulated by Page et al. [1], gives a rank list of importance of pages related to user’s query terms. The hyperlink structure of the web can be viewed as a direct graph and modeled by a Markov chain:

(1) A x = [ α P + ( 1 α ) v e T ] x = x ,

where α ( 0 , 1 ) is the damping factor, P n × n is a column stochastic matrix, e is a column vector of all ones, v is called the personalization or the teleportation vector and set as v = e / n . In particular, A is defined as the Google matrix and is both irreducible and aperiodic, which implies that there exists a unique right nonnegative eigenvector x for (1). For more details about the PageRank algorithm, we refer the reader to [1,2,3,4].

Given the huge size and density of the Google matrices, computing PageRank is faced with the big challenge of computational resources, and only a small set of computational tools can come in handy. The power method was first considered to compute the PageRank for its stable and reliable performances. However, when the largest eigenvalue of matrix A is not separated well from the second one, the power method costs more and works less well. Some accelerated techniques have been proposed to speed up its convergence, such as vector extrapolation [5,6,7,8], Arnoldi-type [9,10], aggregation/disaggregation [11], lumping [12,13], adaptive methods [4,14] and inner-outer (IO) iteration methods [15,16,17,18,19,20].

Recently, researchers have focused on the PageRank’s linear system and the corresponding iterative methods have raised concerns [3,15,16,17,20]. Gleich et al. [15] proposed an IO iteration method, which is implemented by solving a linear system with a lower damping factor and similar algebraic structure to the original one, and Gu et al. [17] put forward an improved algorithm, i.e., the power-inner-outer (PIO) iteration. In this paper, we proposed a variant of the IO(PIO) iteration by applying multi-step power and multi-step splitting to combine with the IO iteration to accelerate the computation of PageRank.

The remainder of this paper is structured as follows. In Section 2, we briefly provide the mechanism of the IO iteration for PageRank problem. In Section 3, we introduce the proposed algorithm and investigate the convergence properties in detail. Numerical examples are given to illustrate the effectiveness of the method in Section 4. Finally, some conclusions are drawn in Section 5.

2 The IO(PIO) iteration

In this section, we begin by briefly introducing the derivation of the IO iteration by Gleich et al. [15]. Recalling e T x = 1 , the eigenvalue problem A x = x can be rewritten as

(2) ( I α P ) x = ( 1 α ) v .

As you know, there is a general agreement that the smaller the damping factor is the easier it is to solve the original PageRank problem. In light of this, Gleich et al. [15] reformulated (2) as

(3) ( I β P ) x = ( α β ) P x + ( 1 α ) v , 0 < β < α .

Then they have the following outer stationary iteration:

(4) ( I β P ) x k + 1 = ( α β ) P x k + ( 1 α ) v , k = 1 , 2 , ,

where x 0 = v is the initial guess, though other choices are possible too. However, solving this linear system with coefficient matrix I β P is still computationally difficult, even though β is small. Therefore, the inner Richardson iteration is used to compute the approximation of x k + 1 .

Setting the right-hand side of (4) as

(5) f = ( α β ) P x k + ( 1 α ) v ,

the inner linear system is defined as

(6) ( I β P ) y = f ,

which is computed by the inner iteration

(7) y j + 1 = β P y j + f , j = 0 , 1 , 2 , , l 1 ,

where y 0 = x k is the initial guess and the computed approximate solution y l is assigned to new x k + 1 . The stopping criteria of the outer iteration and the inner iteration are, respectively, given as

( 1 α ) v ( I α P ) x k + 1 1 < τ

and

f ( I β P ) y j + 1 1 < η ,

where the parameters τ and η are the outer and inner tolerances, respectively.

Based on the above discussion, a basic IO iterative algorithm for PageRank problem has been proposed in [15] and its convergence has also been analyzed therein.

From (2) and by the power method, Gu et al. [17] obtained the Richardson iteration

(8) x k + 1 = α P x k + ( 1 α ) v , k = 0 , 1 , 2 , ,

and then they proposed the PIO iteration as follows:

(9) x k + 1 2 = α P x k + ( 1 α ) v , ( I β P ) x k + 1 = ( α β ) P x k + 1 2 + ( 1 α ) v , k = 0 , 1 , 2 , ,

where x 0 = v is the initial guess and β ( 0 , α ) . The first iteration of (9) is easy to implement. In the second iteration of (9), Gu et al. come up with the IO iteration to get the next approximation x k + 1 . They then proved the superiority of the PIO over the power method and the IO iteration and some convergence properties of the PIO method have been given as follows.

Theorem 2.1

[17] The iteration matrix T ( α , β ) of the PIO iteration (9) is given by

(10) T ( α , β ) = α ( α β ) ( I β P ) 1 P 2 ,

and the modulus of its eigenvalues is bounded by

(11) σ = α ( α β ) ( 1 β ) , α ( 0 , 1 ) , 0 < β < α .

Therefore, the spectral radius satisfies ρ ( T ( α , β ) ) σ < 1 , i.e., the PIO iteration converges to the unique solution x of (2) for any initial vector x 0 .

Theorem 2.2

[17] Suppose the second iteration of the PIO iteration is exact and 0 < α < 1 . Then the PIO iteration converges for any β ( 0 , α ) . And it has

(12) x k + 1 x 1 α ( α β ) ( 1 β ) x k x 1

and

(13) x k + 1 x k 1 α ( α β ) ( 1 β ) x k x k 1 1 .

3 The multi-power and multi-splitting IO iteration for PageRank

3.1 The MPMIO iteration

It is not hard to find that the first step of PIO iteration (9) is the classical power method. In this way, the PIO iteration can be understood as the combination of one-step power method and IO iteration. It inspires us to expand the IO(PIO) iteration to more efficient methods. First, we consider the multi-step power method with the IO iteration. Then on the second iteration of (9), we consider the multi-step splitting iteration. Hence, the corresponding variant of the IO(PIO) algorithm, MPMIO, can be proposed. Section 4 is about the details, and the convergence properties are described in Section 3.2.

Given an initial guess x 0 , we first expand the first step of (9) to the multi-step power as follows:

(14) x k + 1 m + 1 = α P x k + (1 α ) v , x k + 2 m + 1 = α P x k + 1 m + 1 + (1 α ) v , x k + m m + 1 = α P x k + m 1 m + 1 + (1 α ) v ,

where α ( 0 , 1 ) , β ( 0 , α ) , v = e / n and m ( 2 ) is the step of the power method.

Now, by introducing two parameters β 1 and β 2 , β 1 , 2 ( 0 , α ) we expand the second step iteration of (9) to the multi-splitting as follows:

(15) ( I β 1 P ) ω k + 1 = ( α β 1 ) P x k + m m + 1 + ( 1 α ) v , ( I β 2 P ) x k + 1 = ( α β 2 ) P ω k + 1 + ( 1 α ) v .

Taken together, the proposed multi-power and multi-splitting IO iteration can be shown below.

Multi-power and multi-splitting IO iteration. Given an initial guess x 0 . Compute

(16) x k + 1 m + 1 = α P x k + ( 1 α ) v , x k + 2 m + 1 = α P x k + 1 m + 1 + ( 1 α ) v , x k + 1 m + 1 = α P x k + m 1 m + 1 + ( 1 α ) v , ( I β 1 P ) ω k + 1 = ( α β 1 ) P x k + m m + 1 + ( 1 α ) v , ( I β 2 P ) x k + 1 = ( α β 2 ) P ω k + 1 + ( 1 α ) v

until the sequence { x k } k = 0 converges.

The first m-step power iterations of (16) can be implemented easily and the IO iteration can be used for the second stage. Setting the right-hand side of the splitting iteration of (16) as

(17) f 1 = ( α β 1 ) P x k + m m + 1 + ( 1 α ) v ,

we get the first inner iteration

(18) y j + 1 = β 1 y j + f 1 , j = 0 , 1 , 2 , , l 1 .

Then we assign ( α β 2 ) P y j + 1 + ( 1 α ) v to f 2 and get the second inner iteration

(19) y j + 1 = β 2 y j + 1 + f 2 , j = 0 , 1 , 2 , , l 1 ,

where we take y 0 = x k + m m + 1 as the initial guess and assign the computed approximation y l to the new x k + 1 .

Now we switch to the first m-step power method and repeat the procedure until the desired PageRank vector is obtained. For the whole iteration, we stop the outer iteration if

( 1 α ) v ( I α P ) x k + 1 1 < τ

is satisfied, and use

f 2 ( I β 2 P ) y j + 1 1 < η

for the inner iteration stopping criterion. Now the main algorithm of this paper is shown in Algorithm 1.

Algorithm 1

(MPMIO)

Input: P, α , β 1 , β 2 , τ , η , v, m
Output: x
1. x v ;
2. y P x ;
3. while α y + (1 α ) v x 1 τ
4.    for i = 1 : m , do
5.        x α y + ( 1 α ) v ;
6.        y = P x ;
7.    end for
8.    f 1 ( α β 1 ) y + (1 α ) v ;
9.    f f 1 + β 1 y ;
10.    f 2 ( α β 2 ) P f + (1 α ) v ;
11.    repeat
12.        x = f 2 + β 2 y ;
13.        y P x ;
14.    until f 2 + β 2 y x 1 < η ;
15.  end while
16. x α y + (1 α ) v .

Lines 1 and 2 of Algorithm 1 initialize x = v and y = P x . The m -step iterations of (16) are done in lines 4–7. Lines 8–10 are used for the computation of f 2 . The inner iteration is implemented in lines 11–14. We use the repeat-until clause to ensure that at least one inner iteration is performed. To terminate the algorithm, line 3 checks the residual of the outer linear (2). And in line 14, the stopping criterion is examined for the inner iteration. At the end of Algorithm 1, a single power method step is used for the possible benefits as given in [21].

3.2 Convergence analysis of the MPMIO iteration

In this subsection, we devote to the convergence properties of our new algorithm and pay particular attention to its superiority over the power method, IO iteration and PIO iteration, respectively.

Return to (16) and substitute the first iteration x k + 1 m + 1 into the second iteration x k + 2 m + 1 , we have

(20) x k + 2 m + 1 = ( α P ) 2 x k + ( 1 α ) v i = 0 1 ( α P ) i .

Keep going and until the last step of the power method, i.e.,

(21) x k + m m + 1 = ( α P ) m x k + ( 1 α ) v i = 0 m 1 ( α P ) i .

At the same time, by introducing two parameters β 1 and β 2 and from the last two steps of (16), we get

(22) x k + 1 = ( α β 1 ) ( α β 2 ) ( I β 1 P ) 1 ( I β 2 P ) 1 P 2 x k + m m + 1 + ( 1 α ) v ( I β 2 P ) 1 .

Substituting x k + m m + 1 into the above formula, we have

(23) x k + 1 = α m ( α β 1 ) ( α β 2 ) ( I β 1 P ) 1 ( I β 2 P ) 1 P m + 2 x k + (1 α ) ( I β 2 P ) 1 v I + ( α β 1 ) ( α β 2 ) ( I β 1 P ) 1 P 2 i =0 m 1 ( α P ) i .

Hence, the iteration matrix of (16) is

(24) T m ( α , β 1 , β 2 ) = α m ( α β 1 ) ( α β 2 ) ( I β 1 P ) 1 ( I β 2 P ) 1 P m + 2 .

Thus, we obtain the following theoretical results for convergence property of the MPMIO iteration (16).

Theorem 3.1

The iteration matrix T m ( α , β 1 , β 2 ) of the MPMIO iteration (16) is given as (24) and the modulus of its eigenvalues is bounded by

(25) σ m ( β 1 , β 2 ) = α m ( α β 1 ) ( α β 2 ) ( 1 β 1 ) ( 1 β 2 ) , α ( 0 , 1 ) , 0 < β 1 , 2 < α .

Meanwhile, the spectral radius satisfies

(26) ρ ( T m ( α , β 1 , β 2 ) ) σ m ( β 1 , β 2 ) < 1 .

In other words, the MPMIO iteration converges to the unique solution x of the linear system (2) for any initial vector x 0 .

Proof

The first part of Theorem 3.1 has been proved. Suppose λ i is an eigenvalue of P, then

μ i = α m ( α β 1 ) ( α β 2 ) λ i m + 2 ( 1 β 1 λ i ) ( 1 β 2 λ i )

is an eigenvalue of the iteration matrix T m ( α , β 1 , β 2 ) . Since | λ i | 1 , we have

(27) | μ i | = α m ( α β 1 ) ( α β 2 ) λ i m + 2 ( 1 β 1 λ i ) ( 1 β 2 λ i ) α m ( α β 1 ) ( α β 2 ) | λ i | m + 2 ( 1 β 1 | λ i | ) ( 1 β 2 | λ i | ) α m ( α β 1 ) ( α β 2 ) ( 1 β 1 ) ( 1 β 2 ) = σ m ( β 1 , β 2 ) ,

with equality holding for λ 1 = 1 . So (25) is proved and the inequality (26) follows directly from (27).□

Remark 3.1

Back to the right-hand side of (25), and given that α ( 0 , 1 ) , β 1 , 2 ( 0 , α ) , we get

σ m ( β 1 , β 2 ) min i α m ( α β i ) 1 β i = min { σ m ( β 1 ) , σ m ( β 2 ) } ,

where σ m ( β 1 ) and σ m ( β 2 ) are the eigenvalue’s upper boundary of the iteration matrix from single-parameter β IO iteration. In this way, it may show the superiority of the multisplitting IO iteration.

Theorem 3.2

Suppose the last splitting iteration of the MPMIO is solved exactly and α ( 0 , 1 ) . Then for any β 1 , 2 ( 0 , α ) , the MPMIO iteration converges. Furthermore,

(28) x k + 1 x 1 α m ( α β 1 ) ( α β 2 ) ( 1 β 1 ) ( 1 β 2 ) x k x 1

and

(29) x k + 1 x k 1 α m ( α β 1 ) ( α β 2 ) ( 1 β 1 ) ( 1 β 2 ) x k x k 1 1 .

Proof

By definition, we have

(30) x = α m ( α β 1 ) ( α β 2 ) ( I β 1 P ) 1 ( I β 2 P ) 1 P m + 2 x + ( 1 α ) ( I β 2 P ) 1 v I + ( α β 1 ) ( α β 2 ) ( I β 1 P ) 1 P 2 i = 0 m 1 ( α P ) i .

Subtracting (30) from (23), we have

(31) x k + 1 x = T m ( α , β 1 , β 2 ) ( x k x ) = α m ( α β 1 ) ( α β 2 ) ( I β 1 P ) 1 ( I β 2 P ) 1 P m + 2 ( x k x ) ,

where T m ( α , β 1 , β 2 ) is the iteration matrix as (24), and x k , x k + 1 refer to the k-step or k + 1 -step iteration, respectively.

Taking 1-norms and using the triangular inequality, we obtain

(32) x k + 1 x 1 α m ( α β 1 ) ( α β 2 ) ( I β 1 P ) 1 1 ( I β 2 P ) 1 1 P 1 m + 2 x k x 1 .

Since the matrix P is column stochastic, it holds that P 1 = 1 . Meanwhile, since I β 1 P and I β 2 P are diagonally dominant M -matrix, their inverses are nonnegative and it follows that

(33) ( I β 1 P ) 1 1 = 1 1 β 1

and

(34) ( I β 2 P ) 1 1 = 1 1 β 2 .

The inequality of (28) follows from (32), (33) and (34).

Setting (23) for step k, subtracting it from (23) and taking norms in the same way like (32), we can easily derive (29).□

Remark 3.2

Denote the contraction factor by

g α ( β 1 , β 2 ) = α m ( α β 1 ) ( α β 2 ) ( 1 β 1 ) ( 1 β 2 ) ,

it is easy to see that

g α ( β 1 , β 2 ) min i g α ( β i ) = min { g α ( β 1 ) , g α ( β 2 ) } ,

where g α ( β i ) , i = 1 , 2 , is the contraction factor of single-parameter IO iteration.

Thus, we can deduce that the MPMIO iteration may converge faster than that of the single parameter β 1 or β 2 .

Remark 3.3

It is easy to prove that g α ( β 1 , β 2 ) is a monotonically decreasing function related to m and thus show the superiority of our MPMIO iteration over the single-power IO iteration, like PIO.

4 Numerical experiments

In this section, we carry out some numerical examples to test the effectiveness of the MPMIO iteration and compare it with the power method, the IO iteration and the PIO iteration. All the numerical results are obtained by using MATLAB9.7.0(R2019b) on a PC with 3.97 GHz Inter(R)Core(TM)i7 processor with 8GB RAM.

For the sake of justice, we take the same initial guess x 0 = v = e / n for all algorithms, where e = [ 1 , 1 , , 1 ] T . All the stopping criteria are set as α P x k + ( 1 α ) v x k 1 < τ , where τ is specified in experiment description. We choose α = 0.85 , 0.90 , 0.95 and 0.99, and set m = 2 , 3 , 5 , 10 , 20 and 50 . Referring to the analysis and empirical choices by Gleich et al. [15] and Gu et al. [17], we use the choices η = 0.01 , β = 0.5 , and assign different values to β 1 and β 2 around β .

The Web matrices are listed in Table 1, where “nnz” denotes the number of nonzero elements and “avg nnz per row” denotes the average number of nonzero elements per row. For convenience, in all the tables to follow we have abbreviated the power method, the inner-outer iteration method, the PIO iteration method and our MPMIO iteration method as Power, IO, PIO, MPMIO, respectively. We denote by “ite” the iteration counts, “mv” the number of matrix-vector products and by “CPU” the CPU time used in seconds.

Table 1

Web matrices for PageRank problems

Name Size nnz Avg nnz per row
Amazon0505 410,236 3,356,824 19.9
Stanford-Berkeley 683,446 7,583,376 11.1
Web-Google 916,428 5,105,039 5.6
Wikipedia-20051105 1,634,989 19,753,078 12.1
Table 2

Numerical results for Example 1 with m = 5 , β = 0.5 , τ = 10 8 and various value pairs of ( β 1 , β 2 )

Method alpha = 0.85 alpha = 0.90 alpha = 0.95 alpha = 0.99
ite (mv) CPU ite (mv) CPU ite (mv) CPU ite (mv) CPU
Power 81 (81) 0.514029 122 (122) 0.730376 238 (238) 1.430304 1,043 (1,043) 6.436183
IO 77 (84) 0.577852 118 (126) 0.888501 233 (244) 1.753881 1,032 (1,047) 7.334066
PIO 39 (82) 0.492393 60 (125) 0.733401 118 (243) 1.421175 517 (1,044) 5.876933
MPMIO (0.5,0.4) 12 (85) 0.429520 18 (127) 0.630644 34 (240) 1.192104 149 (1,045) 5.381957
MPMIO (0.5,0.5) 12 (85) 0.439055 18 (128) 0.662717 34 (240) 1.221052 148 (1,040) 5.294960
MPMIO (0.6,0.4) 12 (85) 0.433683 18 (127) 0.634136 34 (240) 1.183174 149 (1,045) 5.342428
MPMIO (0.6,0.5) 12 (85) 0.436202 18 (128) 0.635568 34 (240) 1.189355 148 (1,040) 5.312114
MPMIO (0.7,0.4) 12 (85) 0.430336 18 (127) 0.642486 34 (240) 1.188378 149 (1,045) 5.301542
MPMIO (0.7,0.5) 12 (85) 0.440045 18 (128) 0.657614 34 (240) 1.239790 148 (1,040) 5.336923
MPMIO (0.7,0.6) 12 (85) 0.434616 18 (128) 0.658883 34 (241) 1.233188 148 (1,040) 5.310513
MPMIO (0.8,0.5) 12 (85) 0.429087 18 (128) 0.652059 34 (240) 1.215909 148 (1,040) 5.319561
MPMIO (0.8,0.6) 12 (85) 0.430386 18 (128) 0.644962 34 (241) 1.222724 148 (1,040) 5.267550
MPMIO (0.8,0.7) 12 (86) 0.428920 17 (122) 0.620944 34 (241) 1.226266 148 (1,041) 5.345917
Table 3

Numerical results for Example 2 with β = 0.5 , β 1 = 0.6 , β 2 = 0.5 , τ = 10 8 and various values of m

Method alpha = 0.85 alpha = 0.90 alpha = 0.95 alpha = 0.99
ite (mv) CPU ite (mv) CPU ite (mv) CPU ite (mv) CPU
Power 91 (91) 1.025699 138 (138) 1.506846 277 (277) 3.041634 1,341 (1,341) 14.958891
IO 76 (85) 1.116774 115 (126) 1.622780 225 (238) 3.112812 1,024 (1,041) 14.065285
PIO 39 (83) 0.913716 58 (123) 1.331178 114 (236) 2.566159 513 (1,038) 11.338970
MPMIO (m = 2) 22 (91) 0.902490 30 (125) 1.243692 58 (237) 2.350283 257 (1,035) 10.291378
MPMIO (m = 3) 18 (93) 0.890123 24 (124) 1.200560 46 (235) 2.289176 206 (1,037) 10.301479
MPMIO (m = 5) 12 (86) 0.816488 19 (136) 1.291546 33 (235) 2.245291 148 (1,041) 10.003695
MPMIO (m = 10) 8 (97) 0.881731 10 (122) 1.112733 22 (267) 2.454582 86 (1,036) 9.521987
MPMIO (m = 20) 5 (111) 0.979710 7 (155) 1.382064 13 (287) 2.558793 58 (1,279) 11.093444
MPMIO (m = 50) 2 (105) 0.900378 3 (157) 1.359985 6 (313) 2.740545 26 (1,353) 11.957493

Example 1

This example aims at discussing some of the options for parameters β 1 and β 2 . The test matrix is amazon0505 Web matrix. With m = 5 , τ = 10 8 , we run the different methods for PageRank problem with different pairs of values for ( β 1 , β 2 ) . Numerical results are presented in Table 2.

It is easy to see that MPMIO performs the best both in terms of iteration numbers and CPU time. As for the matrix-vector products, they are approximately equal and with the increase in α , MPMIO gradually reflects its advantages. At the same time, we find that there are still different performances with different values of parameters β 1 and β 2 , and it is currently very hard to get the best choices of β 1 and β 2 . Thus, we choose empirically the parameters as β 1 = 0.6 and β 2 = 0.5 in the following experiments.

Example 2

In this example, we examine the performance of the four methods for PageRank with various values of m. The test matrix is the Stanford-Berkeley Web matrix. Numerical results are listed in Table 3.

From Table 3, we find that among the four methods, MPMIO outperforms the other three methods, i.e., Power, IO, PIO, especially in iteration counts and CPU time. Meanwhile, we can find that the performance of MPMIO is sensitive to the choice of m. For example, when m takes a small value MPMIO is outstanding but when m takes a big value, like 20 or 50, the performance of MPMIO is barely satisfactory. Then, based on other similar observations of test matrices, we tend to choose a modest value of m, i.e., m = 5 .

Example 3

In this example, we give a comprehensive investigation into the performance of MPMIO. We choose empirically the parameters as α = 0.99 , β = 0.5 , β 1 = 0.6 , β 2 = 0.5 , m = 5 , η = 0.01 . We test all the matrices in Table 1 to compare the number of matrix-vector products and CPU time required for convergence to three different outer tolerances τ . To state a speedup of an algorithm a over another one b, we use speedup formula as

Speedup = CPU b CPU a .

Numerical results are given in Table 4.

This example shows that our proposed algorithm MPMIO can reduce the number of matrix-vector products obviously and is proved to be efficient. Based on the CPU time, MPMIO iteration performs the best for each test PageRank problem. When τ = 10 8 , MPMIO achieves a speedup from 1.13× to 1.48× over IO and from 1.04× to 1.20× over PIO. Taken together, our proposed MPMIO iteration converges faster than the other three methods, and this has indeed been shown by the convergent curves in Figure 1.

Table 4

Numerical results for Example 3 with α = 0.99 , m = 5 , β = 0.5 , β 1 = 0.6 , β 2 = 0.5 , η = 0.01 . Speedup represents the relative speedup in CPU time of MPMIO over IO and PIO (in brackets)

Tol Web matrix mv CPU
Power IO PIO MPMIO Power IO PIO MPMIO Speedup
10 4 Amazon0505 283 291 288 243 1.906843 1.979214 1.940777 1.559590 1.27 (1.24)
Stanford-Berkeley 473 260 258 220 5.106515 3.266863 2.748681 2.371718 1.38 (1.16)
Web-Google 489 304 303 261 10.132581 6.862155 5.917642 5.195980 1.32 (1.14)
Wikipedia20051105 42 44 43 37 2.890107 3.146707 2.877156 2.810747 1.12 (1.02)
10 6 Amazon0505 652 659 656 561 4.127204 4.772426 4.057435 3.375080 1.41 (1.20)
Stanford-Berkeley 902 631 628 538 10.429152 8.750668 7.169710 6.206567 1.41 (1.16)
Web-Google 942 636 635 549 18.705083 14.338529 12.547242 11.190281 1.28 (1.12)
Wikipedia20051105 390 342 341 295 26.986150 25.628100 23.665184 22.695913 1.13 (1.04)
10 8 Amazon0505 1,043 1,047 1,044 891 6.608148 7.811836 6.310431 5.269707 1.48 (1.20)
Stanford-Berkeley 1,341 1,041 1,038 892 14.471853 13.509255 10.938215 9.651750 1.40 (1.13)
Web-Google 1,399 991 993 867 27.265529 22.284097 19.162797 17.304339 1.29 (1.11)
Wikipedia20051105 847 799 799 691 59.130899 59.895641 55.289397 53.219597 1.13 (1.04)
Figure 1 
               Convergence of the computation for the four Web matrix. 
                     
                        
                        
                           α
                           =
                           0.99
                           ,
                           
                           τ
                           =
                           
                              
                                 10
                              
                              
                                 −
                                 8
                              
                           
                           .
                        
                        \alpha =0.99,\hspace{.5em}\tau ={10}^{-8}.
Figure 1

Convergence of the computation for the four Web matrix. α = 0.99 , τ = 10 8 .

5 Conclusions

In this paper, we have improved the IO iteration for accelerating PageRank computation by introducing multi-step power and multi-step splitting. Our algorithm can be implemented easily, and theoretical results show its efficiency. Numerical experiments on several PageRank problems have indicated that the new algorithm is superior to the power method and the IO iteration methods, IO and PIO. At the same time, we have also realized that the new algorithm is parameter-dependent and appropriate choice of parameters can be made in our experiments. It is meaningful to explore how to determine the optimal parameters for our algorithm and may be included in the future work.

Acknowledgments

This research was supported by the Key Fund Project of Sichuan Provincial Department of Education (17za0003).

References

[1] L. Page, S. Brin, R. Motwani, and T. Winograd, The PageRank Citation Ranking: Bringing Order to the Web, Technical Report, Stanford InfoLab, 1999.Search in Google Scholar

[2] A. N. Langville and C. D. Meyer, Deeper inside PageRank, Internet Math. 1 (2004), no. 3, 335–380, https://doi.org/10.1080/15427951.2004.10129091.10.1080/15427951.2004.10129091Search in Google Scholar

[3] A. N. Langville and C. D. Meyer, Google’s PageRank and Beyond: The Science of Search Engine Rankings, Princeton University Press, Princeton, NJ, 2012.Search in Google Scholar

[4] P. Berkhin, A survey on PageRank computing, Internet Math. 2 (2005), no. 1, 73–120, https://doi.org/10.1080/15427951.2005.10129098.10.1080/15427951.2005.10129098Search in Google Scholar

[5] S. D. Kamvar, T. H. Haveliwala, C. D. Manning, and G. H. Golub, Extrapolation methods for accelerating PageRank computations, in: WWW ‘03 – Proceedings of the 12th International Conference on World Wide Web, Association for Computing Machinery, New York, NY, USA, 2003, pp. 261–270.10.1145/775152.775190Search in Google Scholar

[6] A. Sidi, Vector extrapolation methods with applications to solution of large systems of equations and to PageRank computations, Comput. Math. Appl. 56 (2008), no. 1, 1–24, https://doi.org/10.1016/j.camwa.2007.11.027.10.1016/j.camwa.2007.11.027Search in Google Scholar

[7] B. Y. Pu, T. Z. Huang and C. Wen, A preconditioned and extrapolation-accelerated GMRES method for PageRank, Appl. Math. Lett. 37 (2014), 95–100, https://doi.org/10.1016/j.aml.2014.05.017.10.1016/j.aml.2014.05.017Search in Google Scholar

[8] H. Migallon, V. Migallon, J. A. Palomino, and J. Penades, A heuristic relaxed extrapolated algorithm for accelerating PageRank, Adv. Eng. Softw. 120 (2018), 88–95, https://doi.org/10.1016/j.advengsoft.2016.01.024.10.1016/j.advengsoft.2016.01.024Search in Google Scholar

[9] X. Tan, A new extrapolation method for PageRank computations, J. Comput. Appl. Math. 313 (2017), 383–392, https://doi.org/10.1016/j.cam.2016.08.034.10.1016/j.cam.2016.08.034Search in Google Scholar

[10] G. Wu and Y. Wei, An Arnoldi-extrapolation algorithm for computing PageRank, J. Comput. Appl. Math. 234 (2010), no. 11, 3196–3212, https://doi.org/10.1016/j.cam.2010.02.009.10.1016/j.cam.2010.02.009Search in Google Scholar

[11] H. De Sterck, T. A. Manteuffel, S. F. McCormick, Q. Nguyen, and J. Ruge, Multilevel adaptive aggregation for Markov chains, with application to web ranking, SIAM J. Sci. Comput. 30 (2008), no. 5, 2235–2262, https://doi.org/10.1137/070685142.10.1137/070685142Search in Google Scholar

[12] Y. Lin, X. Shi, and Y. Wei, On computing PageRank via lumping the Google matrix, J. Comput. Appl. Math. 224 (2009), no. 2, 702–708, https://doi.org/10.1016/j.cam.2008.06.003.10.1016/j.cam.2008.06.003Search in Google Scholar

[13] I. R. Mendes and P. B. Vasconcelos, PageRank computation with MAAOR and lumping methods, Math. Comput. Sci. 12 (2018), no. 2, 129–141, https://doi.org/10.1007/s11786-018-0335-7.10.1007/s11786-018-0335-7Search in Google Scholar

[14] S. Kamvar, T. Haveliwala, and G. Golub, Adaptive methods for the computation of PageRank, Linear Algebra Appl. 386 (2004), 51–65, https://doi.org/10.1016/j.laa.2003.12.008.10.1016/j.laa.2003.12.008Search in Google Scholar

[15] D. Gleich, A. Gray, C. Greif, and T. Lau, An inner-outer iteration for computing PageRank, SIAM J. Sci. Comput. 32 (2010), 349–371, https://doi.org/10.1137/080727397.10.1137/080727397Search in Google Scholar

[16] Z. Z. Bai, On convergence of the inner-outer iteration method for computing PageRank, Numer. Algebra Control Optim. 2 (2012), no. 4, 855–862, https://doi.org/10.3934/naco.2012.2.855.10.3934/naco.2012.2.855Search in Google Scholar

[17] C. Gu, F. Xie, and K. Zhang, A two-step matrix splitting iteration for computing PageRank, J. Comput. Appl. Math. 278 (2015), 19–28, https://doi.org/10.1016/j.cam.2014.09.022.10.1016/j.cam.2014.09.022Search in Google Scholar

[18] C. Gu and W. Wang, An Arnoldi-Inout algorithm for computing PageRank problems, J. Comput. Appl. Math. 309 (2017), 219–229, https://doi.org/10.1016/j.cam.2016.05.026.10.1016/j.cam.2016.05.026Search in Google Scholar

[19] Y. J. Xie and C. F. Ma, A relaxed two-step splitting iteration method for computing PageRank, Comp. Appl. Math. 37 (2018), 221–233, https://doi.org/10.1007/S40314-016-0338-4.10.1007/s40314-016-0338-4Search in Google Scholar

[20] Z. Tian, Y. Liu, Y. Zhang, Z. Liu, and M. Tian, The general inner-outer iteration method based on regular splittings for the PageRank problem, Appl. Math. Comput. 356 (2019), 479–501, https://doi.org/10.1016/j.amc.2019.02.066.10.1016/j.amc.2019.02.066Search in Google Scholar

[21] R. S. Wills and I. C. F. Ipsen, Ordinal ranking for Google’s PageRank, SIAM J. Matrix Anal. Appl. 30 (2009), no. 4, 1677–1696, https://doi.org/10.1137/070698129.10.1137/070698129Search in Google Scholar

Received: 2020-06-04
Revised: 2020-11-24
Accepted: 2020-11-24
Published Online: 2020-12-31

© 2020 Bing-Yuan Pu et al., published by De Gruyter

This work is licensed under the Creative Commons Attribution 4.0 International License.

Articles in the same Issue

  1. Regular Articles
  2. Non-occurrence of the Lavrentiev phenomenon for a class of convex nonautonomous Lagrangians
  3. Strong and weak convergence of Ishikawa iterations for best proximity pairs
  4. Curve and surface construction based on the generalized toric-Bernstein basis functions
  5. The non-negative spectrum of a digraph
  6. Bounds on F-index of tricyclic graphs with fixed pendant vertices
  7. Crank-Nicolson orthogonal spline collocation method combined with WSGI difference scheme for the two-dimensional time-fractional diffusion-wave equation
  8. Hardy’s inequalities and integral operators on Herz-Morrey spaces
  9. The 2-pebbling property of squares of paths and Graham’s conjecture
  10. Existence conditions for periodic solutions of second-order neutral delay differential equations with piecewise constant arguments
  11. Orthogonal polynomials for exponential weights x2α(1 – x2)2ρe–2Q(x) on [0, 1)
  12. Rough sets based on fuzzy ideals in distributive lattices
  13. On more general forms of proportional fractional operators
  14. The hyperbolic polygons of type (ϵ, n) and Möbius transformations
  15. Tripled best proximity point in complete metric spaces
  16. Metric completions, the Heine-Borel property, and approachability
  17. Functional identities on upper triangular matrix rings
  18. Uniqueness on entire functions and their nth order exact differences with two shared values
  19. The adaptive finite element method for the Steklov eigenvalue problem in inverse scattering
  20. Existence of a common solution to systems of integral equations via fixed point results
  21. Fixed point results for multivalued mappings of Ćirić type via F-contractions on quasi metric spaces
  22. Some inequalities on the spectral radius of nonnegative tensors
  23. Some results in cone metric spaces with applications in homotopy theory
  24. On the Malcev products of some classes of epigroups, I
  25. Self-injectivity of semigroup algebras
  26. Cauchy matrix and Liouville formula of quaternion impulsive dynamic equations on time scales
  27. On the symmetrized s-divergence
  28. On multivalued Suzuki-type θ-contractions and related applications
  29. Approximation operators based on preconcepts
  30. Two types of hypergeometric degenerate Cauchy numbers
  31. The molecular characterization of anisotropic Herz-type Hardy spaces with two variable exponents
  32. Discussions on the almost 𝒵-contraction
  33. On a predator-prey system interaction under fluctuating water level with nonselective harvesting
  34. On split involutive regular BiHom-Lie superalgebras
  35. Weighted CBMO estimates for commutators of matrix Hausdorff operator on the Heisenberg group
  36. Inverse Sturm-Liouville problem with analytical functions in the boundary condition
  37. The L-ordered L-semihypergroups
  38. Global structure of sign-changing solutions for discrete Dirichlet problems
  39. Analysis of F-contractions in function weighted metric spaces with an application
  40. On finite dual Cayley graphs
  41. Left and right inverse eigenpairs problem with a submatrix constraint for the generalized centrosymmetric matrix
  42. Controllability of fractional stochastic evolution equations with nonlocal conditions and noncompact semigroups
  43. Levinson-type inequalities via new Green functions and Montgomery identity
  44. The core inverse and constrained matrix approximation problem
  45. A pair of equations in unlike powers of primes and powers of 2
  46. Miscellaneous equalities for idempotent matrices with applications
  47. B-maximal commutators, commutators of B-singular integral operators and B-Riesz potentials on B-Morrey spaces
  48. Rate of convergence of uniform transport processes to a Brownian sheet
  49. Curves in the Lorentz-Minkowski plane with curvature depending on their position
  50. Sequential change-point detection in a multinomial logistic regression model
  51. Tiny zero-sum sequences over some special groups
  52. A boundedness result for Marcinkiewicz integral operator
  53. On a functional equation that has the quadratic-multiplicative property
  54. The spectrum generated by s-numbers and pre-quasi normed Orlicz-Cesáro mean sequence spaces
  55. Positive coincidence points for a class of nonlinear operators and their applications to matrix equations
  56. Asymptotic relations for the products of elements of some positive sequences
  57. Jordan {g,h}-derivations on triangular algebras
  58. A systolic inequality with remainder in the real projective plane
  59. A new characterization of L2(p2)
  60. Nonlinear boundary value problems for mixed-type fractional equations and Ulam-Hyers stability
  61. Asymptotic normality and mean consistency of LS estimators in the errors-in-variables model with dependent errors
  62. Some non-commuting solutions of the Yang-Baxter-like matrix equation
  63. General (p,q)-mixed projection bodies
  64. An extension of the method of brackets. Part 2
  65. A new approach in the context of ordered incomplete partial b-metric spaces
  66. Sharper existence and uniqueness results for solutions to fourth-order boundary value problems and elastic beam analysis
  67. Remark on subgroup intersection graph of finite abelian groups
  68. Detectable sensation of a stochastic smoking model
  69. Almost Kenmotsu 3-h-manifolds with transversely Killing-type Ricci operators
  70. Some inequalities for star duality of the radial Blaschke-Minkowski homomorphisms
  71. Results on nonlocal stochastic integro-differential equations driven by a fractional Brownian motion
  72. On surrounding quasi-contractions on non-triangular metric spaces
  73. SEMT valuation and strength of subdivided star of K 1,4
  74. Weak solutions and optimal controls of stochastic fractional reaction-diffusion systems
  75. Gradient estimates for a weighted nonlinear parabolic equation and applications
  76. On the equivalence of three-dimensional differential systems
  77. Free nonunitary Rota-Baxter family algebras and typed leaf-spaced decorated planar rooted forests
  78. The prime and maximal spectra and the reticulation of residuated lattices with applications to De Morgan residuated lattices
  79. Explicit determinantal formula for a class of banded matrices
  80. Dynamics of a diffusive delayed competition and cooperation system
  81. Error term of the mean value theorem for binary Egyptian fractions
  82. The integral part of a nonlinear form with a square, a cube and a biquadrate
  83. Meromorphic solutions of certain nonlinear difference equations
  84. Characterizations for the potential operators on Carleson curves in local generalized Morrey spaces
  85. Some integral curves with a new frame
  86. Meromorphic exact solutions of the (2 + 1)-dimensional generalized Calogero-Bogoyavlenskii-Schiff equation
  87. Towards a homological generalization of the direct summand theorem
  88. A standard form in (some) free fields: How to construct minimal linear representations
  89. On the determination of the number of positive and negative polynomial zeros and their isolation
  90. Perturbation of the one-dimensional time-independent Schrödinger equation with a rectangular potential barrier
  91. Simply connected topological spaces of weighted composition operators
  92. Generalized derivatives and optimization problems for n-dimensional fuzzy-number-valued functions
  93. A study of uniformities on the space of uniformly continuous mappings
  94. The strong nil-cleanness of semigroup rings
  95. On an equivalence between regular ordered Γ-semigroups and regular ordered semigroups
  96. Evolution of the first eigenvalue of the Laplace operator and the p-Laplace operator under a forced mean curvature flow
  97. Noetherian properties in composite generalized power series rings
  98. Inequalities for the generalized trigonometric and hyperbolic functions
  99. Blow-up analyses in nonlocal reaction diffusion equations with time-dependent coefficients under Neumann boundary conditions
  100. A new characterization of a proper type B semigroup
  101. Constructions of pseudorandom binary lattices using cyclotomic classes in finite fields
  102. Estimates of entropy numbers in probabilistic setting
  103. Ramsey numbers of partial order graphs (comparability graphs) and implications in ring theory
  104. S-shaped connected component of positive solutions for second-order discrete Neumann boundary value problems
  105. The logarithmic mean of two convex functionals
  106. A modified Tikhonov regularization method based on Hermite expansion for solving the Cauchy problem of the Laplace equation
  107. Approximation properties of tensor norms and operator ideals for Banach spaces
  108. A multi-power and multi-splitting inner-outer iteration for PageRank computation
  109. The edge-regular complete maps
  110. Ramanujan’s function k(τ)=r(τ)r2(2τ) and its modularity
  111. Finite groups with some weakly pronormal subgroups
  112. A new refinement of Jensen’s inequality with applications in information theory
  113. Skew-symmetric and essentially unitary operators via Berezin symbols
  114. The limit Riemann solutions to nonisentropic Chaplygin Euler equations
  115. On singularities of real algebraic sets and applications to kinematics
  116. Results on analytic functions defined by Laplace-Stieltjes transforms with perfect ϕ-type
  117. New (p, q)-estimates for different types of integral inequalities via (α, m)-convex mappings
  118. Boundary value problems of Hilfer-type fractional integro-differential equations and inclusions with nonlocal integro-multipoint boundary conditions
  119. Boundary layer analysis for a 2-D Keller-Segel model
  120. On some extensions of Gauss’ work and applications
  121. A study on strongly convex hyper S-subposets in hyper S-posets
  122. On the Gevrey ultradifferentiability of weak solutions of an abstract evolution equation with a scalar type spectral operator on the real axis
  123. Special Issue on Graph Theory (GWGT 2019), Part II
  124. On applications of bipartite graph associated with algebraic structures
  125. Further new results on strong resolving partitions for graphs
  126. The second out-neighborhood for local tournaments
  127. On the N-spectrum of oriented graphs
  128. The H-force sets of the graphs satisfying the condition of Ore’s theorem
  129. Bipartite graphs with close domination and k-domination numbers
  130. On the sandpile model of modified wheels II
  131. Connected even factors in k-tree
  132. On triangular matroids induced by n3-configurations
  133. The domination number of round digraphs
  134. Special Issue on Variational/Hemivariational Inequalities
  135. A new blow-up criterion for the Nabc family of Camassa-Holm type equation with both dissipation and dispersion
  136. On the finite approximate controllability for Hilfer fractional evolution systems with nonlocal conditions
  137. On the well-posedness of differential quasi-variational-hemivariational inequalities
  138. An efficient approach for the numerical solution of fifth-order KdV equations
  139. Generalized fractional integral inequalities of Hermite-Hadamard-type for a convex function
  140. Karush-Kuhn-Tucker optimality conditions for a class of robust optimization problems with an interval-valued objective function
  141. An equivalent quasinorm for the Lipschitz space of noncommutative martingales
  142. Optimal control of a viscous generalized θ-type dispersive equation with weak dissipation
  143. Special Issue on Problems, Methods and Applications of Nonlinear analysis
  144. Generalized Picone inequalities and their applications to (p,q)-Laplace equations
  145. Positive solutions for parametric (p(z),q(z))-equations
  146. Revisiting the sub- and super-solution method for the classical radial solutions of the mean curvature equation
  147. (p,Q) systems with critical singular exponential nonlinearities in the Heisenberg group
  148. Quasilinear Dirichlet problems with competing operators and convection
  149. Hyers-Ulam-Rassias stability of (m, n)-Jordan derivations
  150. Special Issue on Evolution Equations, Theory and Applications
  151. Instantaneous blow-up of solutions to the Cauchy problem for the fractional Khokhlov-Zabolotskaya equation
  152. Three classes of decomposable distributions
Downloaded on 10.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/math-2020-0120/html
Scroll to top button