Next Article in Journal
Information-Theoretic Sequential Framework to Elicit Dynamic High-Order Interactions in High-Dimensional Network Processes
Previous Article in Journal
A Comprehensive Review of Image Restoration Research Based on Diffusion Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PageRank of Gluing Networks and Corresponding Markov Chains

1
The Independent Schools Foundation Academy, 26A Tower 2 Bel Air No. 8 Phase 6, Pok Fu Lam, Hong Kong
2
Qiuzhen College, Tsinghua University, Beijing 100084, China
3
Yau Mathematical Sciences Center, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2025, 13(13), 2080; https://doi.org/10.3390/math13132080
Submission received: 10 May 2025 / Revised: 17 June 2025 / Accepted: 20 June 2025 / Published: 24 June 2025
(This article belongs to the Section E: Applied Mathematics)

Abstract

This paper studies Google’s PageRank algorithm. By an innovative application of the method of gluing Markov chains, we study the properties of Markov chains and extend their applicability by accounting for the damping factor and the personalization vector. Many properties of Markov chains related to spectrums and eigenvectors of the transition matrix, including the stationary distribution, periodicity, and persistent and transient states, will be investigated as well as part of the gluing process. Using the gluing formula, it is possible to decompose a large network into some sub-networks, compute their PageRank separably and glue them together. The computational workload can be reduced.
MSC:
60J10; 05C81

1. Introduction

In this section, we give a brief introduction to Markov chains and their applications. And Google PageRank, as an important application of Markov chains, is the main object with which we deal in this paper. Then we introduce our main ideas, such as gluing Markov chains to study Google PageRank, and show the advantage of these ideas compared to traditional methods.

1.1. Recent Results for Markov Chains

Markov chains were introduced by Andrey Markov in 1906 [1] to model systems transitioning between states with probabilities. Markov chains appear as the behind-the-scenes numerical structure in a huge amount of fields; thus, they have their application in a huge amount of areas, such as seismology, computer networks, epidemiology, and social dynamics.
Markov chains have been effectively applied in seismology to model earthquake occurrences and assess seismic hazards. Bountzis et al. [2] used the MAP method to analyze the temporal evolution of earthquakes in the Corinth Gulf, Greece, revealing high burstiness in the eastern subarea and demonstrating that MAPs outperform traditional models like the Gamma distribution. Votsi et al. [3] applied a discrete-time Markov model to the Hellenic Subduction Zone, estimating key indicators such as hitting times and failure rates and providing probabilistic seismic hazard results. Ünal et al. [4] combined k-means clustering with Markov chains to study earthquake sequencing in the Aegean Graben system of Turkey, identifying stationary distributions and mean passage times for different clusters. These studies highlight the versatility of Markov Chains in capturing the complex dynamics of earthquake occurrences and improving hazard predictions.
The application of Markov chains in modeling the spread of computer viruses within networks has been greatly explored in recent studies. Amador and Artalejo in [5] explore the stochastic SIRS model with warning signals, providing additional insights into the model’s behavior and characteristics. Then Amador [6] investigates a stochastic SEIQS model for computer viruses using continuous-time Markov chains, analyzing key characteristics such as stationary distributions, extinction and hazard times, and transient behaviors of susceptible, exposed, infectious, and quarantined computers. Papageorgiou and Vasiliadis [7] introduced a discrete-time non-homogeneous Markov system with state capacities, related to an SIQS model for computer networks. Their research highlights the importance of quarantine capacity and the impact of infection rates on network sustainability.
Markov chains are actively applied in epidemiology to improve the understanding and prediction of epidemic dynamics. Economou and Lopez–Herrero [8] studied the deterministic SIS model in a Markovian random environment, describing the influence of environmental changes on infection and recovery rates. Papageorgiou and Tsaklidis [9] further extended the stochastic SIRD model, focusing on the computation of stochastic properties such as the alert time and the infection and mortality times of tagged individuals. Papageorgiou [10] developed a stochastic SIRD model with imperfect immunity, investigating various stochastic properties like the extinction time of epidemics and the infection and mortality times of individuals. Papageorgiou et al. [11] proposed a novel method integrating the Extended Kalman Filter with recursive algorithms to estimate stochastic epidemic descriptors, such as the duration of outbreaks and the total number of infections and deaths. These studies collectively demonstrate the versatility and effectiveness of Markov chain models in analyzing epidemic phenomena and guiding public health interventions.
Markov chains are also useful to model social dynamics, studying how opinions or behaviors evolve through interactions in a network. They help analyze the impact of influence and network structure on collective behaviors. Kindermann R P, Snell J L. [12] gave a detailed introduction to this area. Recently, Bolzern et al. [13] showed that influence intensity affects opinion variance but not the mean under unbiased conditions. Similarly, Papageorgiou [14] used Markov chains to model the spread of radicalization, introducing new stochastic descriptors to better assess its severity.

1.2. Introduction for Google PageRank

Markov chains have wide applications, as the above shows. However, when it comes to one of the most influential and widespread applications of Markov chain, it is Google PageRank.
To understand how Markov chains appear on web pages, the CNN and Yahoo homepages are excellent examples. These pages are divided into distinct sections or branches, such as sports, economy, politics, health, etc. It is observed that there are numerous internal hyperlinks within each section, while the links connecting different sections are relatively small. Therefore, there should be a hidden algorithm designed to assign weights to each hyperlink, thereby determining their frequency of appearance.
Using this idea, Google founders Larry Page and Sergey Brin discovered a novel idea using graph theory, formulating the well-known Google PageRank [15]: A network can be seen as a directed graph. There should also be some numerical information on the edge to measure the importance or frequency of the corresponding link. In a word, these data form Markov chains over graphs [16].
More precisely, a network can be realized as a directed graph; more specifically, it is a graph with a finite vertex set, where each vertex represents a website and a directed edge from a to b if and only if there is a hyperlink from a to b. If there is a web surfer that starts from some page and goes to other pages in the same network via hyperlinks, this behavior is called a random walk on a directed graph. After a long time, it is hoped that the probability distribution converges to a limit depending on the initial distribution. The limiting distribution will be called the PageRank, and this value is usually used to rank the pages.
We will describe this by the following example:
Mathematics 13 02080 i001
Here, we study four websites and their hyperlinks.
  • If the web surfer is at 1 , it has 2 outgoing hyperlinks to 2 and 4 , so in the next step the web surfer has probability 1 2 to go to 2 , and 1 2 to go to 4 .
  • Similarly, from 2, the web surfer has only 1 hyperlink to 1 , so there is a probability of 1 of going to 1, and so on...
In conclusion, if it has a limiting probability distribution, we can use x 1 , x 2 , x 3 , x 4 to express their probability. Then it satisfies the following linear equations:
x 2 + 1 3 x 4 = x 1 1 2 x 1 + 1 2 x 3 + 1 3 x 4 = x 2 1 3 x 4 = x 3 1 2 x 1 + 1 2 x 3 = x 4 ,
Solving this equation with the extra condition x 1 + x 2 + x 3 + x 4 = 1 , we have the probability distribution
( x 1 , x 2 , x 3 , x 4 ) = 5 13 , 4 13 , 1 13 , 3 13 ,
and this solution is unique. Then we have the rank of the pages 1 > 2 > 4 > 3 .
Remark 1. 
However, not all examples work using the algorithm above to yield a unique solution.
There are several good materials for the history and details of the Google PageRank, and we list some of them following. Rogers [17] explained with many examples how the weight can be designed to get the response faster. There are also many calculations and examples helping to understand how the PageRank works. Avrachenkov and Litvak [18] gave a more detailed discussion of the effect of newly created links and the extent to which a page can control its PageRank. They employed asymptotic analysis and mathematical modeling to study the impact of new links on Google PageRank. And they used Markov chain theory and matrix operations to derive conditions under which new links can increase the PageRank of a webpage and its neighbors, showing that linking within a relevant web community benefits PageRank, while irrelevant links can penalize the webpage and its community. Langville and Meyer [19] provided a comprehensive overview of the PageRank algorithm and its mathematical foundations. They used linear algebra, Markov chains, and numerical analysis to explain how PageRank is calculated and how it influences search engine rankings. Their book also shows many different algorithms that Google has applied for its research engines. Joshi and Patel [20] studied updates over time for the Google PageRank. They analyzed the impact of various factors such as backlinks, website age, content relevancy, traffic, keyword occurrence, and domain authority on PageRank through empirical observations and case studies. Their work shows that backlinks from high-authority websites significantly boost PageRank, while content quality and relevancy are crucial for maintaining high rankings. Frequent updates to the algorithm, such as Panda and Penguin, emphasize the importance of avoiding spammy practices and focusing on user experience.

1.3. Structure and Main Results of This Paper

In this paper, we introduce an approach to analyzing PageRank by employing the concept of gluing Markov chains. This method involves combining multiple Markov chains into a single chain by identifying certain states, allowing us to study the properties of the resulting chain and its relation to the original chains. This approach is particularly useful for understanding how the PageRank of a large network can be decomposed into smaller sub-networks, computed separately, and then recombined. This not only reduces the computational workload but also provides insights into the structural properties of the network.
The structure of this article is linear; we will first introduce the preliminaries, then establish the following results in order, and finally leave a section for discussion.
Our main results consist of the following parts:
After we introduce the concept of gluing Markov chains, we study the stationary distribution of the θ gluing Markov chain Theorem 2.
Then we study the performance of some other properties of Markov chains after gluing, including the number of irreducible components Theorem 3 and the periodicity Theorem 4.
Lastly, we consider the network with the personalization vector and study how to glue the personalization vector as we glue the underlying Markov chains Theorem 7.
The idea of gluing Markov chains has its advantages compared to other results on computing PageRank; we provide some detailed discussion as follows.
Traditional methods for computing PageRank, as introduced by Page and Brin [15], rely on iterative algorithms such as the Power Iteration method. Their methods compute the PageRank values directly from the entire web graph, treating it as a single Markov chain. While effective, these methods can be computationally intensive, especially for large web graphs. For example, the power iteration method requires multiple iterations over the entire graph to converge to the stationary distribution, which can be time-consuming and resource-intensive. Our gluing approach decomposes the large network into smaller sub-networks, computes their PageRank values separately, and then recombines them. This reduces the computational complexity by allowing parallel processing of sub-networks and minimizing the number of iterations required for convergence. Additionally, our method provides a deeper understanding of the structural properties of the network by analyzing the interactions between sub-networks.
Personalized PageRank, introduced by Haveliwala [21], extends the traditional PageRank algorithm by incorporating user preferences through a personalization vector. This allows for more tailored search results and has been widely adopted in personalized search and recommendation systems. Personalized PageRank modifies the probability transition matrix by adding a damping factor and a personalization vector, ensuring convergence to a unique stationary distribution. While personalized PageRank focuses on modifying the transition matrix to incorporate user preferences, our gluing approach focuses on the structural decomposition of the network. Our method can be combined with personalized PageRank to further enhance the efficiency and relevance of search results. By decomposing the network, we can apply personalized PageRank to smaller sub-networks, reducing computational overhead and improving scalability.
Distributed computing approaches, such as those proposed by Kamvar et al. [22], leverage parallel processing to compute PageRank values more efficiently. These methods divide the web graph into smaller partitions and compute PageRank values in parallel across multiple processors or nodes. While these methods significantly reduce computation time, they require sophisticated distributed systems infrastructure and coordination mechanisms. Our gluing approach is inherently parallelizable, as it decomposes the network into smaller sub-networks that can be processed independently. The idea of decomposing networks in their approach is similar to ours. However, unlike distributed PageRank methods that require synchronization and communication between nodes, our method allows for independent computation of PageRank values for each sub-network. This reduces the complexity of distributed system design and improves the overall efficiency of the computation.
Approximate methods, such as those proposed by Andersen et al. [23], aim to reduce the computational complexity of PageRank by approximating the stationary distribution. These methods often use sampling techniques or truncated iterations to estimate PageRank values, trading off accuracy for computational efficiency. Our gluing approach provides an exact method for computing PageRank values by leveraging the structural properties of the network. While approximate methods can be faster, they may not always provide accurate results. Our method ensures precise computation of PageRank values by decomposing and recombining the results, maintaining accuracy while improving efficiency.

2. Preliminaries

In this section, we give the necessary preliminaries. Most of them are basic definitions and propositions on Markov chains.

2.1. Basic Definitions

We basically work with Markov chains and graphs. Now we give some explanation for their definitions.
Definition 1 
(Probability Distribution Vector and Stochastic Matrix ([24], Page 2)). A probability distribution vector of n states is a vector x = ( x 1 , , x n ) T where for 1 i n , we have 0 x i 1 and i = 1 n x i = 1 . An n × n matrix is called a stochastic matrix if its columns are probability distribution vectors.
Definition 2 
(Finite Markov Chain ([24], Page 2)). A discrete-time finite Markov chain M = ( S , A ) consists of the following data: Firstly, the set of states S = { x 1 , x n } , where the elements x i are functions from N to R 0 such that the vector x ( t ) = ( x 1 ( t ) , , x n ( t ) ) T is a probability distribution vector of n states, t N ; secondly, an n × n stochastic matrix A such that x ( t + 1 ) = A · x ( t ) , t N .
The vector x ( 0 ) is called the initial distribution, and A is called the probability transition matrix of the Markov chain.
Each element a i , j of probability transition matrix A reflects the probability of transition from x j to x i . And the stochastic condition is necessary to ensure that x t are probability distributions for t 1 .
For a more generalized definition, see [25].
Now we fix some notation for general directed graphs G = ( V , E ) ([26], Page 1). Usually, we will assume that there is no self-loop in the graph. There is an associated state vector x corresponding to G. And we can construct a transition matrix A for such a graph (for precise definitions, see [24], Page 8).
So we can construct a Markov chain from the graph, denoted by M = ( S , A ) (known as random walk on graphs, see [24], Page 8).
Definition 3 
(Stationary Distribution ([27], Page 21)). The stationary distribution (or equilibrium state) of a Markov chain is a probability distribution vector of n-states. x satisfies x = Ax .
In the model of a network, we also call the corresponding stationary distribution the PageRank vector, meaning that each time we transfer importance, each page gives away the same importance it receives.
We also deal with undirected graphs. An undirected graph G = ( V , E ) is a directed graph such that for any pair of vertices i , j V the number of edges from i to j is the same as the number of j to i.
In the study of Markov chains, we would like to know the behavior of x ( t ) when t gets large and how it depends on x ( 0 ) . There are also many examples and applications of Markov Chain other than PageRank.
Remark 2. 
There exists a Markov chain that has a unique stationary distribution, but the sequence of probability vector x ( t ) : = A t x ( 0 ) does not converge as t + . For example, consider the Markov chain with x ( 0 ) = ( 1 , 0 , 0 , 0 ) T and the transition matrix is of the following form:
0 1 0 0 0 0 1 0 1 0 0 0 0 0 0 0 .
It is easy to verify the above claim for such a Markov chain.

2.2. Spectrum of A

Both the uniqueness of the stationary distribution and the convergence of x ( t ) depend on the eigenvalues of A . In this section, we recall the related results. The most important is Perron–Frobenius theorem. It is a deep theorem leading to fruitful consequences; one can get more in [28].
Theorem 1 
(Perron-Frobenius [29]). For a matrix A with all entries non-negative such that A m has all positive entries for some m, there exists a positive eigenvalue λ 0 of multiplicity one such that any absolute value of other eigenvalues is strictly less than λ 0 . And there exists a left eigenvector of λ 0 that has all positive entries.
Proof. 
See [30] or Page 125 of [29]. □
For a general stochastic matrix, one can derive the following classical results.
Corollary 1 
([29], Page 126). For an n × n stochastic matrix A , let λ 1 , , λ n be all the eigenvalues of A ; then we have the following properties:
(1)
The absolute value of eigenvalue | λ i | 1 .
(2)
The number 1 is an eigenvalue of the matrix.
(3)
For any eigenvalue satisfying | λ i | = 1 , then λ i is a root of unity. Especially when the elements in A are all positive, then such λ i = 1 and the algebraic multiplicity of eigenvalue λ = 1 is 1.
From the above corollary, we can also see that for general stochastic matrices, the algebraic multiplicity of eigenvalue λ = 1 is the same as the geometric multiplicity.
Proposition 1. 
When all entries in stochastic matrix A are positive, then A k x converges as k approaches infinity for any probability distribution vector x . The limit does not depend on the choice of x .
Note that the eigenvalue λ = 1 has multiplicity 1, and other eigenvalues have norm strictly less than 1 by Theorem 1. The proposition follows from the Jordan canonical form and direct computations.
However, by the example given in Remark 2, this does not necessarily work when not every entry is positive. Motivated by this, we add the damping factor to make A k x converge.

2.3. Damping Factor

In PageRank, there is a modification of the Markov Chain by adding personal interests and a factor 0 < d < 1 into the discussion to guarantee convergence. This happens for two reasons:
  • In real life, the web surfer has probability ( 1 d ) to follow the hyperlinks to visit the next website and also probability d to jump to any of the pages according to the surfer’s personal interests. The probability for the random jump to the i-th website is denoted by p i .
  • It makes all entries of the transition matrix positive, guaranteeing convergence, which will be explained as follows.
Call the probability distribution vector p = ( p 1 , , p n ) T the personalization vector, which gives a weighting of which pages the web surfer is likely to randomly jump to and has all positive entries. Then let d be the damping factor. For each iteration, we have x ( t + 1 ) = ( 1 d ) A x ( t ) + d · p . Observe this converges as we end up with a matrix with fully positive entries.
x ( t + 1 ) = [ ( 1 d ) A + d ( p ( 1 , , 1 ) ) ] x ( t ) .
Let C = ( 1 d ) A + d p ( 1 , , 1 ) . Then C is a stochastic matrix, and all the entries of C are positive. This ensures the convergence of x ( t ) and the uniqueness of the limiting stationary distribution. The absolute value of the second eigenvalue has an upper bound of 1 d , see [31]. This gives a convergence rate of x ( t ) . In real life, Google usually uses a damping factor of 0.15 to achieve a good PageRank result for customers.
Remark 3. 
It is important to emphasize that the personalization vector p is a probability distribution vector. On one hand, it reflects the probability for the random jump; thus, it should be automatically normalized. On the other hand, C = [ ( 1 d ) A + d ( p ( 1 , , 1 ) ) ] x ( t ) being a stochastic matrix requires the condition that p is a probability distribution vector.

2.4. Basic Properties of States

In this section, we recall some properties of states in Markov chain M = ( S , A ) .
Definition 4 
(Closed; [25] Page 16). A subset T of S is called closed if it satisfies that for j T and i S T , then the transition probability a i , j = 0 .
A closed subset T can be viewed as a Markov chain by inheriting the original transformations a i , j from M.
Definition 5 
(Irreducible Closed Subset; [27] Page 39). Let a i , j m be the i , j -th element (rather than a power of a i , j ) in A m : = A × × A (for m-times). Two states i , j intercommunicate if m 1 , m 2 such that a i , j m 1 > 0 and a j , i m 2 > 0 , where a i , j m is the ( i , j ) -th element of A m .
A closed subset T S in the Markov chain M is called irreducible if and only if all pairs of states in T intercommunicate.
When a transition matrix has eigenvalue 1 with multiplicity, it means the network can be split into multiple irreducible closed subsets. The dimension of the eigenspace would cause the existence of multiple linearly independent eigenvectors, which would be the PageRank vectors of the irreducible subsets. In particular, we have the following proposition.
Proposition 2. 
For any Markov chain, then there is a decomposition of S = S 1 S k S where each S j is an irreducible closed subset and S is the set not containing any irreducible closed subsets. This splitting is unique.
Proof. 
See Page 126 of [29] □
Observe the decomposition yields a transition matrix of the form
A = B 1 0 0 R 1 0 B 2 0 R 2 0 0 B k R k 0 0 0 Q ,
where B i denotes the probability transition matrix of the Markov chain M i , which corresponds to S i . Each B i contributes exactly one multiplicity for eigenvalue 1. And Q reflects the probability transition from S to itself, and R i reflects the probability transition from S to S i .
Definition 6 
(Page 53 [27]).
  • For each state, it is persistent if there is a probability 1 of a process starting from this state to eventually return to it.
  • If a state is not persistent, it is transient. That is, if some importance leaves this state, it has a positive probability of not returning to it at all.
For instance, each state in a connected undirected graph is persistent.
In Proposition 2, all the states in S are exactly the transient states. It is suggested to consult [32] to obtain more properties of states for readers who are interested.

3. Gluing of Markov Chains and Stationary Distributions

3.1. Definitions of Gluing

We would like to study the behavior of gluing graphs for two major reasons:
  • In abstract graph theory, it is very natural to consider the gluing of graphs so that a complicated graph can sometimes be decomposed or cut into smaller pieces.
  • It has real-world applications to sites that are hubs that lead to other disparate networks. Consider, for instance, a new website. From the main page, one can find many different hyperlinked pages—politics, sports, etc. However, these form disparate networks that might not hyperlink to each other at all. In that sense, the only reason they are not separate irreducible networks is due to the presence of a main hub that they intercommunicate with. If we see the main hub as a page in each of those networks that were glued together, we can calculate the individual PageRanks pre-gluing and then put them together to find the final PageRank, saving computational power. We can also do the same with personalization vectors and see how the local personalization vectors affect the global personalization vector.
First, we consider an example.
Example 1. 
Consider the gluing of two networks given by 1 , 2 , 3 and 3 , 4 , respectively:
Mathematics 13 02080 i002
The first network has PageRank 2 9 , 4 9 , 3 9 and the second has 1 2 , 1 2 . The new network after gluing has PageRank 2 15 , 4 15 , 6 15 , 3 15 . The weighting is 9 15 to 6 15 . This leads us to consider whether we can develop the following proposition for gluing two Markov chains with equal weight, which is also the induction basis for the case of more general gluing processes:
Note that the networks and PageRank vectors here are actually Markov chains and their stationary distributions. Our arguments following fit for arbitrary Markov chains and their stationary distributions. So, from now on, we will replace networks and PageRank vectors by Markov chains and their stationary distributions.
We define the gluing of two graphs as follows.
Definition 7. 
For i = 1 , 2 , let G i = ( V i , E i ) be a directed graph with vertex set V i and edge E i . Let v 0 i be a fixed vertex in V i . The gluing of G 1 and G 2 along v 0 1 and v 0 2 is a graph G = ( V , E ) with vertex set
V = V 1 V 2 / v 0 1 v 0 2 ,
which is the union of V 1 and V 2 quotient by the relation v 0 1 v 0 2 . The quotient point is denoted by v 0 = v 0 1 = v 0 2 . The edge set is E = E 1 E 2 .
We have the following relations among probability transition matrices for random walks on these three graphs.
Proposition 3. 
Let B = ( b v w ) v , w V 1 be the probability transition matrix for random walk on graph G 1 and C = ( c v w ) v , w V 2 be the probability transition matrix for graph G 2 . Then the probability transition matrix A = ( a v w ) v , w V for random walk on graph G has the following form:
a v w = b v w if v V 1 , w V 1 { v 0 } c v w if v V 2 , w V 2 { v 0 } θ 1 b v w if v V 1 { v 0 } , w = v 0 θ 2 c v w if v V 2 { v 0 } , w = v 0 θ 1 b v 0 v 0 + θ 2 c v 0 v 0 if v = w = v 0 0 otherwise ,
for some weights θ 1 + θ 2 = 1 . More explicitly, let g i ( v 0 i ) be the number of outgoing edges from v 0 i in G i , here
θ 1 = g 1 ( v 0 1 ) g 1 ( v 0 1 ) + g 2 ( v 0 2 ) and θ 2 = g 2 ( v 0 2 ) g 1 ( v 0 1 ) + g 2 ( v 0 2 ) .
Here, the weight θ i , i = 1 , 2 can be seen as the probability of transition between two networks, which is accessible from the quotient point. Note that the gluing process can be generalized to fit for arbitrary finitely many graphs and their transition matrices. And note that a Markov chain is a combination of graphs and a more generalized probability transition matrix (that is, not the natural random walk transition matrix). So we can easily generalize to the concept of gluing finitely many Markov chains by analogue the process of gluing graphs and gluing transition matrix.
Definition 8 
(Gluing of Markov Chains). For integers i = 1 , 2 , , k , let M i = ( S i , A i ) be k Markov chains, and v 0 i be a fixed point in S i . Assume θ = ( θ 1 , θ 2 , , θ k ) , 0 < θ i < 1 , and
i = 1 k θ i = 1 .
The θ-gluing of M i along v 0 i is the following Markov chain M = ( S , A ) . The state set
S = i = 1 k S i / v 0 1 v 0 2 v 0 k
is the disjoint union of S i quotient by relation v 0 1 v 0 2 v 0 k and denote this quotient point by v 0 = v 0 1 = v 0 2 = = v 0 k . Let A i = ( a i , v w ) v , w S i . Then the transition matrix A is defined by
a v w = a i , v w if v S i , w S i { v 0 } θ i a i , v w if v S i { v 0 } , w = v 0 i = 1 k θ i a i , v 0 v 0 if v = w = v 0 0 otherwise .

3.2. Theorem of Gluing Stationary Distributions

Proposition 4. 
Under the notation and assumption of the definition above, gluing two Markov chains M 1 = ( S 1 , A 1 ) and M 2 = ( S 2 , A 2 ) at vertices v 0 1 S 1 and v 0 2 S 2 with weight ( θ 1 , θ 2 ) , we obtain a new Markov chain M = ( S , A ) . Let y = ( y v ) v V 1 , z = ( z v ) v V 2 be stationary distributions of M 1 , M 2 , respectively. Then there is a stationary distribution x = ( x v ) v V of the new glued Markov chain with the following stationary distribution
x v = α y v if v V 1 { v 0 1 } β z v if v V 2 { v 0 2 } α y v 0 1 + β z v 0 2 if v = v 0 1 = v 0 2 ,
where
α = θ 1 z v 0 2 θ 1 z v 0 2 + θ 2 y v 0 1 = θ 1 1 y v 0 1 θ 1 1 y v 0 1 + θ 2 1 z v 0 2 β = θ 2 y v 0 1 θ 1 z v 0 2 + θ 2 y v 0 1 = θ 2 1 z v 0 2 θ 1 1 y v 0 1 + θ 2 1 z v 0 2 .
Proof. 
Consider the stationary equation Ax = x . For each row, we have equation x v = w a v w x w . Now we have three cases for index v, v V 1 { v 0 1 } , v V 2 { v 0 2 } and v = v 0 1 = v 0 2 .
In the first case, v V 1 { v 0 1 } . Then we have
α y v = x v = w x a v w x w = w V 1 { v 0 1 } a v w x w + a v v 0 1 x v 0 1 + w V 2 { v 0 2 } a v w x w = w V 1 { v 0 1 } b v w · α y w + a v v 0 1 x v 0 1 = w V 1 { v 0 1 } b v w · α y w + θ 1 b v v 0 ( α y v 0 2 + β z v 0 2 ) = α w V 1 b v w · y w ( 1 θ 1 ) b v v 0 α y v 0 2 + θ 1 b v v 0 β z v 0 2 = α y v θ 2 b v v 0 α y v 0 2 + θ 1 b v v 0 β z v 0 2 ,
yielding
θ 2 b v v 0 α y v 0 2 = θ 1 b v v 0 β z v 0 2 θ 2 α y v 0 2 = θ 1 β z v 0 2 α β = θ 1 z v 0 2 θ 2 y v 0 1 .
Analogously, we can find the same α to β ratio in the second case v V 2 { v 0 2 } . For the third case, take the α = β ratios given from the previous two parts of the ansatz. Then
α y v 0 1 + β z v 0 2 = w V 1 { v 0 1 } θ 1 a v 0 1 w · x w + w V 2 { v 0 2 } θ 2 a v 0 2 w x w α w V 1 { v 0 1 } a v 0 1 w y w + β w V 2 { v 0 2 } a v 0 2 w z w = w V 1 { v 0 1 } θ 1 a v 0 1 w · x w + w V 2 { v 0 2 } θ 2 a v 0 2 w x w ,
for which the substitution x w = α y w for w V 1 { v 0 1 } and x w = β z w for w V 2 { v 0 2 } works. □
The above proposition can be inductively generalized to the following theorem, which describes the stationary distribution after a general weighted gluing.
Theorem 2 
(Gluing Stationary Distributions). Let M = ( S , A ) be a θ-gluing of Markov chains M i in Definition 8. Let x i = ( x v i ) v S i be a stationary distribution of M i and assume x v 0 i 0 . Denote by
α i : = θ i x v 0 i and α : = i = 1 k α i .
Then there is a x = ( x v ) v S of M
x v = α i α x v i if v S i { v 0 i } i = 1 k α i α x v 0 i if v = v 0 .
Proof. 
Suppose that this holds until the gluing of k Markov chains. Take the gluing of the k + 1 Markov chains. First, we glue the k Markov chains together; we then glue them with the ( k + 1 ) th network.
More explicitly, from θ = ( θ 1 , , θ k + 1 ) , we define a system of new weights θ ˜ = ( θ 1 ˜ , , θ k ˜ ) with
θ i ˜ = θ i i = 1 k θ i
The θ ˜ -gluing of M i = ( S i , A i ) gives a Markov chain M ˜ . Then we can see that the ( 1 θ k + 1 , θ k + 1 ) gluing of M ˜ with M k + 1 is exactly the θ -gluing of M 1 , , M k + 1 .
The proof of Proposition 4 does not use the explicit form of θ in terms of outgoing edges. So it still works for a θ -gluing of two Markov chains. So we obtain the required identity based on the gluing formula for equilibrium vectors of two Markov chains. □
Example 2. 
Let us use the following example to illustrate that the gluing method does save computing power.
We use the code provided by the following website: http://www.pagerank.dk/, to perform the calculations of the examples.
Consider the following two Markov chains coming from random walks on graphs with corresponding probability transition matrices of the following form, respectively:
1 3 1 2 0 0 1 3 0 1 1 2 1 3 1 2 0 0 0 0 0 1 2 and 0 0 0 0 1 3 1 2 0 0 1 2 0 0 1 2 1 1 2 1 3 1 2 1 2 0 0 0 0 0 0 0 1 3 .
We glue the last state of the first Markov chain and the first state of the second Markov chain together and we use the gluing weight as Proposition 3. Then the gluing probability transition matrix is the following:
1 3 1 2 0 0 0 0 0 0 1 3 0 1 1 4 0 0 0 0 1 3 1 2 0 0 0 0 0 0 0 0 0 1 4 0 0 0 0 0 0 0 0 0 0 0 1 3 0 0 0 1 4 0 0 1 2 0 0 0 0 0 1 2 1 1 2 1 3 0 0 0 1 4 1 2 0 0 0 0 0 0 0 0 0 0 1 3 .
Firstly, we compute the stationary state of the gluing Markov chain directly. Secondly, we compute the stationary state for the two Markov chains and glue them together by the formula in Theorem 2. We compare the iterations at which the stationary distribution converges to three decimal places in above two cases.
The first case, that is, the classical method to compute the stationary distribution, spends 47 iterations, as the following table shows.
Classical Method
Iteration01020304050
state 10.1250.0520.0140.0040.0010.000
state 20.1250.0380.0100.0030.0010.000
state 30.1250.0140.0040.0010.0000.000
state 40.1250.0380.0090.0030.0010.000
state 50.1250.0000.0000.0000.0000.000
state 60.1250.0000.0000.0000.0000.000
state 70.1250.0000.0000.0000.0000.000
state 80.1250.8580.9620.9900.9971.000
While the second case, which is our gluing method, spends 14 iterations:
Gluing Method
Iteration01020304050
state 10.1250.0000.0000.0000.0000.000
state 20.1250.0000.0000.0000.0000.000
state 30.1250.0000.0000.0000.0000.000
state 40.1250.0020.0000.0000.0000.000
state 50.1250.0000.0000.0000.0000.000
state 60.1250.0010.0000.0000.0000.000
state 70.1250.0000.0000.0000.0000.000
state 80.1250.9961.0001.0001.0001.000
From the simple example, we see that the gluing method does improve the computational resource. We believe the advantage becomes more obvious as the gluing Markov chain becomes of larger and larger size.

3.3. Example of Undirected Graphs

Recall that an undirected graph G = ( V , E ) is a directed graph such that for any pair of vertices v , w V , the number of edges from v to w is the same as the number from w to v. In this section, we consider the Markov chains corresponding to random walks on undirected graphs; that is, the probability transition matrices are of the following form:
a v w v , w V = k v w g ( w ) v , w V .
Here k v w is the number of edges from vertices v to w. And g ( w ) is the valence of w [33], which is the number of edges that start from w.
For undirected graphs G = ( V , E ) , there is a canonical stationary distribution given by the valence of each vertex as follows. Let e be the number of total edges in E. Then the vector
x = g ( v ) e v V
automatically satisfies the equilibrium or PageRank equation. This is a direct computation; see page 4 of [33] for the case that k v w = 0 or 1 . We call this the canonical stationary distribution or the valence PageRank. Since the valence of the vertex is additive under gluing, we have the following observation.
Lemma 1. 
For a glued graph G of two undirected graphs G 1 , G 2 without loops, the addition of valence remains the valence for G.
In Theorem 2, we only allow 1 state in each Markov chain to be glued. But for two undirected graphs G 1 , G 2 , we can allow arbitrarily many pairs of states to be glued.
Actually, the key obstruction for gluing arbitrary states in the general case is that the α and β in Proposition 4 is dependent on the choice of pair of states to be glued, where α and β can be seen as the distribution of importance of these two Markov chains. So for arbitrarily many pairs of states, we do not know which pair to choose to determine α and β . However, when G 1 and G 2 are undirected graphs, as we will see, such α and β are independent of the choice of gluing pair. Therefore, they can be glued for arbitrarily many pairs of points in a natural way.
Let P be the set of pairs of vertices ( v 1 , v 2 ) such that v i V i , for i = 1 , 2 . We assume that for any v V 1 , there exists at most 1 element w V 2 such that ( v , w ) P , and symmetrically, we make the same assumption for V 2 . We glue G 1 and G 2 along P to get an undirected graph G. Let L i be the subset of V i such that v V i , v appears as the i -th coordinate of some element in P, for i = 1 , 2 . Let L be a subset of vertex set V of G such that v L comes from the gluing of some pairs in P.
We define the gluing probability transition matrix A : = ( a v w ) v , w V as follows:
a v w = k v w g ( w ) if v V i , w V i L i θ i w k v w i g ( w i ) if v V i L i , w L θ 1 w k v 1 w 1 g ( w 1 ) + θ 2 w k v 2 w 2 g ( w 2 ) if v , w L 0 otherwise i = 1 , 2 .
Note that for each w L , there exists a unique pair ( w 1 , w 2 ) P gluing to w, now θ w i : = g ( v i ) g ( v 1 ) + g ( v 2 ) , analogous to the weight construct in Definition 7. One can check that matrix A is nothing but the probability transition matrix for random walk on G.
Now, analogous to Proposition 4, we compute the corresponding α and β for each ( θ w 1 , θ w 2 ) . Then they are independent of the choice of w L . Indeed, let e i be the number of edges of G i . Then we have α = e 1 e 1 + e 2 and β = e 2 e 1 + e 2 . Then we claim that the following formula analogous to Proposition holds:
Proposition 5. 
For G 1 = ( V 1 , E 1 ) and G 2 = ( V 2 , E 2 ) be two undirected graphs.
Let P, L i for i = 1 , 2 , α, and β be defined as above, and let G = ( V , E ) be the gluing undirected graph of G 1 and G 2 along P with L gluing points. And denote x i = g ( v i ) e i v i V i by the canonical stationary distribution of random walk on G i .
If we define the natural gluing stochastic vector x = ( x v ) v V as follows:
x v = α g ( v ) e 1 if v V 1 L 1 β g ( v ) e 2 if v V 2 L 2 α g ( v 1 ) e 1 + β g ( v 2 ) e 2 if v L , glued from ( v 1 , v 2 ) ,
then x is the canonical stationary distribution for random walk on G.
Indeed, x = ( x v ) v V equals g ( v ) ˜ e 1 + e 2 v V , where g ( v ) ˜ is the valence of v in the gluing graph, which is the sum of valences of v in G 1 and G 2 by lemma 1 if v is one of the gluing states. Now, G is undirected again and A is the probability transition matrix of the random walk on G, then x is a canonical stationary distribution of G as we introduced before.
That is, Proposition can be generalized to arbitrary many gluings for undirected graphs.
The gluing graph of undirected graphs without loops remains undirected without loops. By induction based on the above argument, a similar result to Theorem holds for gluing arbitrarily many states for undirected graphs.

4. Other Properties of Gluing Markov Chains

4.1. Number of Irreducible Components

Theorem 3. 
When two Markov chains M 1 = ( S 1 , A 1 ) and M 2 = ( S 2 , A 2 ) with k 1 and k 2 irreducible closed subsets, respectively, are θ glued to a Markov chain M, the number of irreducible closed subsets in the glued network is either k 1 + k 2 or k 1 + k 2 1 .
Proof. 
Let a S 1 and b S 2 be glued together to s 0 . Let S 1 be split into irreducible closed subsets P i for 1 i k 1 and P , while M 2 is split into Q i for 1 i k 2 and Q .
  • If the glued points are part of P i and Q j for some i , j , then P i and Q j now form one irreducible and closed network, and all other irreducible and closed networks in S are still irreducible and closed, as other networks in S do not intercommunicate with any point in P i and thus not any T . This yields a total of k 1 + k 2 1 closed irreducible networks.
  • If the glued points are part of P i and Q or P and Q j , then P i and Q j are still irreducible and closed networks, yielding k 1 + k 2 .
  • If the glued points are part of P , Q , then all closed irreducible networks are still closed and irreducible, and the glued point is not closed and irreducible, yielding k 1 + k 2 .
Proposition 6. 
For a finite Markov chain M = ( S , A ) , the multiplicity of eigenvalue 1 in the transition matrix A is the same as the number of irreducible closed subsets in S.
Proof. 
See Page 126 of [29]. □
Corollary 2. 
The multiplicity of eigenvalue 1 after the gluing above is either k 1 + k 2 or k 1 + k 2 1 .
Proof. 
The multiplicity of eigenvalue 1 is the number of irreducible closed subsets, of which there are either k 1 + k 2 or k 1 + k 2 1 .

4.2. Periodicity

Consider a gluing of graphs 1, 2, 3, 4 and 4, 5, 6, 7, 8, 9.
Mathematics 13 02080 i003
Before gluing, each graph appears to cycle 4 and 6 values, respectively, but when they are glued together, some importance starting at 4 can return after either 4 or 6 cycles but still must always return at a number that is a multiple of gcd ( 4 , 6 ) = 2 . We can formalize this definition using periodicity:
Definition 9 
(Periodicity). Denote a i j n as the i j -th entry in A n : = A × × A , where A is the transition matrix. We say a node j has period d if
  • If a j j n > 0 then n = m d for some constant d ,
  • d is the maximum possible integer with the above property.
The periods of all j S for an irreducible Markov chain M = ( S , A ) are the same and called the period of M.
Theorem 4. 
When we glue points with periods b , c , respectively, the new period of the glued point (and thus the glued graph) is gcd ( b , c ) .
Proof. 
The statement of periodicity tells us that if we take an infinitely small node of importance and distribute it across the Markov chain any way we want, it will always reach the gluing point after a number of moves that is a multiple of the period. Starting from v 0 , we can move to a new point in network V 1 , and the next time we reach v 0 , we must have traveled m steps where m 1 mod a by definition. The other option is to move to a new point in network V 2 , and the next time we reach v 0 , we must have traveled n steps where n 1 mod a . Take any path that takes this infinitely small node of importance and split it into “groups,” where the ith group denotes the number of steps needed to go from reaching v 0 the i 1 -th time to the i-th time, and we start at 0 times. Summing these groups gives us the amount of steps needed, which is a linear combination of a , b . This must be 0 mod gcd ( a , b ) , so gcd ( a , b ) satisfies the first criterion for period. To satisfy the second, take configurations traveling in networks V 1 and V 2 only, respectively, to show that the period must divide both a , b . □
Applying this to the previous example, the period is gcd ( 4 , 6 ) = 2 . Based on the period of each state, we have the following corollary.
Corollary 3. 
The period of a θ-gluing of k irreducible Markov chains is the greatest common divisor of the periods of the k Markov chains.
Remark 4. 
In the case that gcd ( b , c ) = 1 , the gluing Markov chain M becomes aperiodic. Then any eigenvalue of M, whose absolute value is 1, is actually 1.
If M is further assumed to be irreducible. Then any eigenvalue is either 1, which has multiplicity 1, or a complex number with absolute value strictly less than 1. Then by direct computation the same as 1 for any initial distribution, the Markov chain convergence, and the limit is independent of choice of the initial distribution.

4.3. Common Eigenvalues and Spectrum

From the proof of gluing PageRank vectors in Theorem 4, we can also consider eigenvalues other than 1 for the gluing matrices.
Theorem 5. 
Use the notation from Definition 7 and the assumption in Proposition 3. Suppose matrix B has eigenvalue λ and matrix C has eigenvalue μ. Then if λ = μ , then it is also an eigenvalue of A .
The proof is the same weighted combination of eigenvectors as Theorem 4. This is consistent with the result for periods of gluing states. We first need a theorem that relates eigenvalues with modulus one to periodicity.
Theorem 6 
([29]). Let A be the transition matrix for an irreducible Markov chain with period d; then all the d-th roots of unity are eigenvalues of A with algebraic multiplicity 1. These are all the eigenvalues of A with an absolute value of 1.
When we glue k irreducible Markov chains with periods d 1 , , d k together, then the common eigenvalues with absolute value 1 for these transition matrices are exactly d-th roots of unity, where d is the greatest common divisor of d 1 , , d k . Based on these observations, we can make a conjecture about the spectrum of gluing Markov chains.
Conjecture 1. 
Glue Markov chains M i = ( S i , A i ) to a Markov chain M = ( S , A ) , then all the eigenvalues of A lie in the convex hull of the eigenvalues of all A i on the complex plane.

5. Personalization Vector Gluing

In this section, we revert to using the terms “network” and “PageRank vector” instead of “Markov chain” and “stationary distribution” to be consistent with terms such as “personalization vector” and “damping factor.”
Now we discuss the PageRank of the gluing of networks together with the damping factor and personalization vector. In the real application of PageRank, the method of damping factors and personalization vectors are usually included to guarantee convergence and provide information based on the web surfer’s personal interests. In Section 2.3, we have seen that the Markov chain with a damping factor model has a strictly positive transition matrix A ˜ . When the two networks are glued together, we have two main differences from θ -gluing of Markov chains. Firstly, the modified transition matrices A ˜ = ( 1 d ) A + d p ( 1 , 1 , , 1 ) are not glued according to the θ -gluing formula. Secondly, usual Markov chains may have many states such that the self-transition probabilities are zero (for example, random walk on a graph without loop), but A ˜ also has positive diagonal entries.
We say the PageRank vector x is for the website M with damping factor d and personalization vector p if it is the solution of A ˜ x = x with A ˜ explained above.
Furthermore, we need a choice of personalization vector for the glued network based on the personalization vectors for both networks M 1 = ( S 1 , A 1 ) and M 2 = ( S 2 , A 2 ) . Let p i = ( p v i ) v V i be the personalization vector for G i . Based on the gluing formula in Theorem 4, we propose a choice of personalization vectors in the following theorem. Let y = ( y v ) v V 1 be the PageRank vector for M 1 with damping factor d and personalization vector p 1 , and let z = ( z v ) v V 2 be the PageRank vector for M 2 with damping factor d and personalization vector p 2 . Under the same notation as Theorem 4, we have the following result.
Theorem 7. 
If the personalization vector is chosen to be p = ( p v ) v V and
p v = α p v 1 if v V 1 , v v 0 1 β p v 2 if v V 2 , v v 0 2 α p v 0 1 + β p v 0 2 if v = v 0 = v 0 1 = v 0 2 ,
with the same α , β as Proposition 4. Then there is a PageRank vector x = ( x v ) v V for the glued website M = ( S , A ) with damping factor d and personalization vector p as follows:
x v = α y v if v V 1 { v 0 1 } β z v if v V 2 { v 0 2 } α y v 0 + β z v 0 if v = v 0 = v 0 1 = v 0 2 .
Proof. 
We need to verify the equilibrium
x v = d j V a v j x j + ( 1 d ) p v .
Take the ansatz
x v = α y i if v V 1 , v v 0 1 β z i if v V 2 , v v 0 2 α y v 0 + β z v 0 if v = v 0 1 = v 0 2 ,
we know
y = d By + ( 1 d ) p 1 z = d Cz + ( 1 d ) p 2 . or y v = d j V 1 b v j y j + ( 1 d ) p v 1 z v = d j V 2 c v j z j + ( 1 d ) p v 2 .
Now, we would like to prove Equation (1) for all v V . If v V 1 and v v 0 1 , we are looking for
α y v = d j V 1 , j v 0 a v j α y j + d · b v v 0 θ 1 ( α y v 0 1 + β z v 0 2 ) + ( 1 d ) p v α y v d j V 1 , j h a v j y j = d · b v v 0 1 θ 1 ( α y v 0 1 + β z v 0 2 ) + ( 1 d ) p v α y v d j V 1 , j h b v j y j = d · b v v 0 1 θ 1 ( α y v 0 1 + β z v 0 2 ) + ( 1 d ) p v ,
and we have
y v = d j V 1 b v j y j + ( 1 d ) p v 1 y v = d v V 1 j h b v j y j + d b v v 0 1 y v 0 1 + ( 1 d ) p v 1 y v d j V 1 , j h b v j y j = d b v v 0 1 y v 0 1 + ( 1 d ) p v 1 α y v d j V 1 , j h b v j y j = α ( d b v v 0 1 y v 0 1 + ( 1 d ) p v 1 ) .
Therefore, it remains to prove
α d b v v 0 1 y v 0 1 + ( 1 d ) p v 1 = d · b v v 0 1 θ 1 ( α y v 0 1 + β z v 0 2 ) + ( 1 d ) p v α ( 1 d ) p v 1 ( 1 d ) p v = d · b v v 0 1 θ 1 ( α y v 0 1 + β z v 0 2 ) α ( d b v v 0 1 y v 0 1 ) α ( 1 d ) p v 1 ( 1 d ) p v = d · b v v 0 1 ( α ( θ 1 1 ) y v 0 1 + β θ 1 z v 0 2 ) α ( 1 d ) p v 1 ( 1 d ) p v = d · b v v 0 1 α θ 2 y v 0 1 + β θ 1 z v 0 2 .
If v 0 1 does not link to v, then b v v 0 1 = 0 , leaving p v = α p v 1 . If v 0 1 does link to v and p v = α p v , then the LHS is 0, and we have
d ( α θ 2 y v 0 1 ) = d ( β θ 1 z v 0 2 ) α θ 2 y v 0 1 = β θ 1 z v 0 2 α β = θ 1 θ 2 · z v 0 2 y v 0 1 .
Next, suppose v V 2 and v v 0 2 . This gives p v = β p v 2 and the same α β ratio as earlier. Finally, suppose v = v 0 = v 0 1 = v 0 2 . Then we have x v 0 1 = x v 0 2 = α y v 0 1 + β z v 0 2 . We would like to have
x v 0 1 = d j V 1 and V 2 a v 0 1 j + ( 1 d ) p v 0 1 ,
or splitting this into two graphs,
x v 0 1 = d j V 1 α b v 0 1 j y j + j V 2 β c v 0 2 j z j + ( 1 d ) p v 0 1 .
Simplifying, the desired equality to be proved is
α ( 1 d ) p v 0 2 1 + β ( 1 d ) p v 0 2 2 = ( 1 d ) p v 0 1 ,
or
v v 0 1 = α p v 0 1 1 + β p v 0 2 2 .
Now we may take:
α β = θ 1 θ 2 · z v 0 2 y v 0 1
and let
p v = α p v 1 if v V 1 , v v 0 1 β p v 2 if v V 2 , v v 0 2 α p v 0 1 1 + β p v 0 2 2 if i = v 0 1 = v 0 2 .
We need v i = 1 , while we already know p v 1 = p v 2 = 1 . This implies α + β = 1 . As α β = θ 1 θ 2 · z v 0 2 y v 0 1 , we have:
α = θ 1 z v 0 2 θ 1 z v 0 2 + θ 2 y v 0 1 β = θ 2 y v 0 1 θ 1 z v 0 2 + θ 2 y v 0 1 .
Theorem 7 proposes a method of gluing personalization vectors. It would be useful to test this proposal in real life. For instance, we could investigate the gluing of personalization vectors for a web surfer on different branches of a news website and test whether the gluing gives a good PageRank recommendation or matches with the usual total personalization vector.

6. Discussion and Future Directions

In this paper, we mainly studied the properties of PageRank vectors when networks are glued together at some vertices. We further generalized this concept to glue finitely many Markov chains and studied various properties of Markov chains under the gluing construction. Those properties are closely related to the eigenvalues of transition matrices.
The novelty of this paper lies in the application of the “gluing” method to Markov chains, specifically in the context of the PageRank formula in Theorem 2. By merging multiple Markov chains into a single chain through the identification of certain states, we are able to study the properties of the resulting chain and its relation to the original chains. This approach allows for the decomposition of a large network into smaller sub-networks, which can be analyzed separately and then recombined. This significantly reduces computational workload by enabling parallel processing and minimizing iterations required for convergence. Actually, from Example 2, one can see the gluing method is about 4 times faster than the classical method even in a very simple case. On the other hand, the gluing method also provides deeper insights into the structural properties of the network. The gluing method also has good compatibility with damping factors and personalization vectors, further enhancing its applicability and efficiency in real-world applications such as search algorithms on the web.
The limitations of this paper primarily stem from the complexities and uncertainties associated with gluing arbitrarily many states in Markov chains. We only establish the gluing formula for gluing 1 state for each Markov chain Theorem 2. However, the PageRank formula for gluing arbitrarily many states is not straightforward to establish. This is because the relative importance for stationary distributions of different Markov chains in the gluing process (such as α and β in Proposition 3.5) is dependent on the choice of gluing states. When multiple states are involved, it becomes unclear which specific pair of states should be used to determine these coefficients. Moreover, the relationships between the glued states themselves add another layer of complexity, making the situation difficult to manage and analyze. However, as demonstrated in the paper, for random walks on undirected graphs, the PageRank formula holds true even when gluing an arbitrary number of states Section 3.3. This highlights a specific case where the challenges of multiple states gluing are successfully addressed, but it also underscores the broader difficulties encountered in more general situations.
There are still many directions we would like to explore in the future about the gluing of Markov chains. For example, the hitting time or cover time represents how long a web surfer needs to reach a certain website or cover all the network. It would be interesting to know how those quantities change under the θ -gluing model. While it would be hard to find the exact formula similar to our stationary distributions, it is still possible to find some bounds or estimates in terms of the hitting time or cover time before gluing. For instance, see ([24], Proposition 10.29) for gluing two identical graphs at a common vertex.
We also would like to investigate more applications of θ -gluing of Markov chains in real life other than PageRank, such as genetics (see [34], Chapter 7.4) or learning theory based on Markov chains (see [34] Chapter 7.5).

Author Contributions

Conceptualization, X.B.H. and C.Y.; methodology, X.B.H.; software, X.B.H. and S.W.; validation, X.B.H., S.W. and C.Y.; formal analysis, X.B.H. and S.W.; investigation, X.B.H.; resources, X.B.H. and C.Y.; data curation, X.B.H.; writing—original draft preparation, X.B.H.; writing—review and editing, S.W. and C.Y.; visualization, X.B.H.; supervision, C.Y.; project administration, C.Y.; funding acquisition, C.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Tsinghua University Talent Recruitment Start-Up Program.

Data Availability Statement

No new data were created or analyzed in this study.

Acknowledgments

The first author is also deeply thankful to Tom Dunne, whose guidance during the MYP Personal Project laid the foundation and inspired the pursuit of this research. Additionally, the first author extends his appreciation to Nicholas Forde and Anastasia Stich for their unwavering support, flexibility, and encouragement in fostering his mathematical endeavors.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Markov, A.A. Rasprostranenie zakona bol’shih chisel na velichiny, zavisyaschie drug ot druga. Izv.-Fiz.-Mat. Obs. Pri Kazan. Univ. 1906, 15, 18. [Google Scholar]
  2. Bountzis, P.; Papadimitriou, E.; Tsaklidis, G. Estimating the earthquake occurrence rates in Corinth Gulf (Greece) through Markovian arrival process modeling. J. Appl. Stat. 2019, 46, 995–1020. [Google Scholar] [CrossRef]
  3. Votsi, I.; Tsaklidis, G.; Limnios, N.; Papadimitriou, E.; Vallianatos, F. A Markov model for seismic hazard analysis along the Hellenic subduction Zone (Greece). Bull. Geol. Soc. Greece 2013, 47, 1376. [Google Scholar] [CrossRef]
  4. Ünal, C.; Özel, G.; Eroglu Azak, T. A Markov chain approach for earthquake sequencing in the Aegean Graben system of Turkey. Earth Sci. Inform. 2023, 16, 1227–1239. [Google Scholar] [CrossRef]
  5. Amador, J.; Artalejo, J.R. Stochastic modeling of computer virus spreading with warning signals. J. Frankl. Inst. 2013, 350, 1112–1138. [Google Scholar] [CrossRef]
  6. Amador, J. The SEIQS stochastic epidemic model with external source of infection. Appl. Math. Model. 2016, 40, 8352–8365. [Google Scholar] [CrossRef]
  7. Papageorgiou, V.E.; Vasiliadis, G. Transient analysis of a SIQS model with state capacities using a non-homogeneous Markov system. J. Frankl. Inst. 2025, 362, 107347. [Google Scholar] [CrossRef]
  8. Economou, A.; Lopez-Herrero, M.J. The deterministic SIS epidemic model in a Markovian random environment. J. Math. Biol. 2016, 73, 91–121. [Google Scholar] [CrossRef] [PubMed]
  9. Papageorgiou, V.E.; Tsaklidis, G. A stochastic SIRD model with imperfect immunity for the evaluation of epidemics. Appl. Math. Model. 2023, 124, 768–790. [Google Scholar] [CrossRef]
  10. Papageorgiou, V.E. Novel stochastic descriptors of a Markovian SIRD model for the assessment of the severity behind epidemic outbreaks. J. Frankl. Inst. 2024, 361, 107022. [Google Scholar] [CrossRef]
  11. Papageorgiou, V.E.; Vasiliadis, G.; Tsaklidis, G. A new method for the estimation of stochastic epidemic descriptors reinforced by Kalman-based dynamic parameter estimation. Application to mpox data. Math. Biosci. 2025, 381, 109365. [Google Scholar] [CrossRef] [PubMed]
  12. Kindermann, R.P.; Snell, J.L. On the relation between Markov random fields and social networks. J. Math. Sociol. 1980, 7, 1–13. [Google Scholar] [CrossRef]
  13. Bolzern, P.; Colaneri, P.; De Nicolao, G. Opinion influence and evolution in social networks: A Markovian agents model. Automatica 2019, 100, 219–230. [Google Scholar] [CrossRef]
  14. Papageorgiou, V.E. Estimating the prevalence of terrorism under control policies. A statistical modelling approach. Appl. Math. Model. 2025, 137, 115642. [Google Scholar] [CrossRef]
  15. Page, L.; Brin, S. The anatomy of a large-scale hypertextual web search engine. Siam Rev. Netw. ISDN Syst. 1998, 30, 1–7. [Google Scholar]
  16. Page, L.; Brin, S.; Motwani, R.; Winograd, T. The PageRank Citation Ranking: Bringing Order to the Web. Technical Report, Stanford Infolab. 1999. Available online: http://ilpubs.stanford.edu:8090/422/?utm_campaign=Technical%20SEO%20Weekly&utm_medium=email&utm_source=Revue%20newsletter (accessed on 23 June 2020).
  17. Ian Rogers. Understanding Google Page Rank. 2002. Available online: https://ianrogers.uk/google-page-rank/ (accessed on 21 June 2020).
  18. Avrachenkov, K.; Litvak, N. The effect of new links on Google pagerank. Stoch. Model. 2006, 22, 319–331. [Google Scholar] [CrossRef]
  19. Langville, A.N.; Meyer, C.D. Google’s PageRank and Beyond: The Science of Search Engine Rankings; Princeton University Press: Princeton, NJ, USA, 2006. [Google Scholar]
  20. Joshi, A.; Patel, P. Google page rank algorithm and it’s updates. Int. J. Manag. Technol. Eng. 2018, 8, 1108–1117. [Google Scholar]
  21. Haveliwala, T.H. Topic-sensitive PageRank. In Proceedings of the 11th International Conference on World Wide Web (WWW ’02), Honolulu, HI, USA, 7–11 May 2002; Association for Computing Machinery: New York, NY, USA, 2002; pp. 517–526. [Google Scholar]
  22. Kamvar, S.; Haveliwala, T.; Golub, G. Adaptive methods for the computation of PageRank. Linear Algebra Its Appl. 2004, 386, 51–65. [Google Scholar] [CrossRef]
  23. Andersen, R.; Chung, F.; Lang, K. Local graph partitioning using pagerank vectors. In Proceedings of the 2006 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS’06), Berkeley, CA, USA, 21–24 October 2006; pp. 475–486. [Google Scholar]
  24. Levin, D.A.; Peres, Y. Markov Chains and Mixing Times, 1st ed.; American Mathematical Soc.: Providence, RI, USA, 2017; Volume 107. [Google Scholar]
  25. Chung, K.L. Markov Chains, 1st ed.; Springer-Verlag: New York, NY, USA, 1967. [Google Scholar]
  26. Bondy, J.A.; Murty, U.S.R. Graph Theory with Applications, 1st ed.; Macmillan: London, UK, 1976; Volume 290. [Google Scholar]
  27. Ursula Porod. Dynamics of Markov Chains for Undergraduates. 2021. Available online: https://www.math.northwestern.edu/documents/book-markov-chains.pdf (accessed on 24 June 2020).
  28. Pillai, S.U.; Suel, T.; Cha, S. The perron-frobenius theorem: Some of its applications. IEEE Signal Process. Mag. 2005, 22, 62–75. [Google Scholar] [CrossRef]
  29. Isaacson, D.L.; Madsen, R.W. Markov Chains Theory and Applications, 1st ed.; John Wiley and Sons: Hoboken, NJ, USA, 1976. [Google Scholar]
  30. MacCluer, C.R. The many proofs and applications of Perron’s theorem. Siam Rev. 2000, 42, 487–498. [Google Scholar] [CrossRef]
  31. Haveliwala, T.; Kamvar, S. The Second Eigenvalue of the Google Matrix. Stanford University Technical Report. 2003. Available online: http://ilpubs.stanford.edu:8090/582/ (accessed on 22 June 2020).
  32. Meyn, S.P.; Tweedie, R.L. Markov Chains and Stochastic Stability, 1st ed.; Springer Science & Business Media: London, UK, 2012. [Google Scholar]
  33. Lovász, L. Random walks on graphs. In Combinatorics, Paul Erdös Is Eighty; Bolyai Society Mathematical Studies; János Bolyai Mathematical Society: Keszthely, Hungary, 1993; Volume 2, pp. 1–46. [Google Scholar]
  34. Kemeny, J.G.; Snell, J.L. Finite Markov Chains, 1st ed.; Van Nostrand: Toronto, ON, Canada, 1960. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, X.B.; Wang, S.; Yu, C. PageRank of Gluing Networks and Corresponding Markov Chains. Mathematics 2025, 13, 2080. https://doi.org/10.3390/math13132080

AMA Style

Han XB, Wang S, Yu C. PageRank of Gluing Networks and Corresponding Markov Chains. Mathematics. 2025; 13(13):2080. https://doi.org/10.3390/math13132080

Chicago/Turabian Style

Han, Xuqian Ben, Shihao Wang, and Chenglong Yu. 2025. "PageRank of Gluing Networks and Corresponding Markov Chains" Mathematics 13, no. 13: 2080. https://doi.org/10.3390/math13132080

APA Style

Han, X. B., Wang, S., & Yu, C. (2025). PageRank of Gluing Networks and Corresponding Markov Chains. Mathematics, 13(13), 2080. https://doi.org/10.3390/math13132080

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop