Google matrix

from Wikipedia, the free encyclopedia
Excerpt from the Google matrix of English-language Wikipedia articles (2009)

The Google matrix is a square matrix that is created when the PageRank algorithm is constructed. Since it is often very large (with many millions of rows and columns), the numerical and algebraic properties of this matrix are of great importance for the quick and exact determination of the PageRanks.

definition

The normalized Google matrix of a network or directed graph with nodes is the real matrix:

The individual components of the Google matrix are defined as follows:

  • The link matrix is the adjacency matrix of the examined graph, standardized line by line :
where is the exit degree of the node , i.e. the number of edges that leave the node .
  • The vector is component-wise defined as
It therefore contains a one if and only if the initial degree of a side or a node is zero. These nodes are also called dangling nodes . There are several methods of treating these lumps in the literature, the most common being the one discussed here.
  • is a real number between and called the damping factor
  • is a one- vector of length , i.e. a vector that has only ones as entries. So the matrix is exactly the one matrix .

properties

PageRank

For the calculation of the PageRanks one is particularly interested in the existence and multiplicity of left eigenvectors of the matrix . These correspond exactly to the usual eigenvectors of the matrix for the eigenvalue . If one interprets the eigenvalue problem

as a calculation of the stationary distribution of a Markov chain , the vector is a stochastic vector consisting of the PageRanks . This reduces the eigenvector problem to the linear system of equations

.

In order to be able to solve this linear system of equations efficiently, the question of the regularity of the matrix and its condition number arises .

Norms

Both the matrix and the matrix are generally only substochastic . If you add both, you get a row-stochastic matrix , since the non-zero rows of the matrices complement each other. Since it is also line stochastic (strictly speaking, even double stochastic ) and only convex combinations are formed by the damping parameter (with regard to which the stochastic matrices are closed), the Google matrix is ​​also a line stochastic matrix. This applies to the line sum norm of the Google matrix

and thus also for the column sum norm of the transposed

.

Eigenvectors and Eigenvalues

The existence of an eigenvector from to the eigenvalue follows directly from the fact that the matrix is ​​a stochastic matrix. That there is even the largest positive eigenvalue, for which there is a simple, strictly positive eigenvector, follows from Perron-Frobenius' theorem , where it holds. It is important here that the introduction of the damping parameter guarantees the positivity of the matrix and thus the solvability of the eigenvalue problem.

Furthermore it can be shown that eigenvalues ​​apply to all other eigenvalues. The separation of the eigenvalues ​​is only determined by the damping parameter. This guarantees a good speed of convergence for many of the numerical methods for calculating eigenvalues, such as the power method , as long as the damping factor is not chosen too close . Usually .

Regularity and stamina

There

holds, the Neumann series provides the invertibility of the matrix

.

Thus the problem can be solved as a linear system of equations. At the same time also applies to the norm of the inverse

and thus the estimate for the condition number

.

Thus, only the choice of the damping parameter is responsible for the condition and should not be too close to .

Numerical calculation of the eigenvector

The largest eigenvector of the Google matrix is ​​usually determined approximately using the power method. Based on a starting approximation, the matrix-vector product of the Google matrix is ​​formed with the current approximation of the eigenvector in each iteration step . In every iteration step is therefore

to calculate. If the starting approximation is a stochastic vector, then every subsequent approximation vector is also stochastic. Since the eigenvalues ​​of the Google matrix are well separated, a slow convergence speed of the power method is excluded.

The special structure of the Google matrix can be used for the calculation. The link matrix is usually extremely sparse , that is, almost all of its entries are zero. As a result, it can on the one hand be saved in a very space-saving manner and, on the other hand, be multiplied very efficiently by a vector. The vector is also usually sparsely populated, which means that the term can also be calculated very quickly.

example

The directed graph treated in the example

If one looks at the directed graph with 8 nodes on the right as an example, nodes 5 and 6 are dangling nodes. Then the row-by-row normalized adjacency matrix is

and the vector

.

Then with the above construction and a damping parameter of

The eigenvector of for eigenvalue 1 is then

.

So nodes 7 and 8 have the highest PageRanks (0.2825 and 0.2654) and nodes 1 and 6 the lowest (0.0675 each). The second eigenvalue is , so the above estimate is sharp. Furthermore is the condition number

,

so this estimate is also sharp.

Individual evidence

  1. Deeper Inside Page Rank Amy N. Langville and Carl D. Meyer. Retrieved August 30, 2013.
  2. TH Haveliwala and SD Kamvar: The Second Eigenvalue of the Google matrix. Technical Report, Stanford University, 2003. Retrieved August 30, 2013.

literature