Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Perron–Frobenius theorem
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
{{Short description|Theory in linear algebra}} In [[matrix theory]], the '''Perron–Frobenius theorem''', proved by {{harvs|txt|authorlink=Oskar Perron|first=Oskar|last= Perron|year=1907}} and {{harvs|txt|authorlink=Georg Frobenius|first=Georg |last=Frobenius|year=1912}}, asserts that a [[real square matrix]] with positive entries has a unique [[eigenvalue]] of largest magnitude and that eigenvalue is real. The corresponding [[eigenvector]] can be chosen to have strictly positive components, and also asserts a similar statement for certain classes of [[nonnegative matrices]]. This theorem has important applications to probability theory ([[ergodicity]] of [[Markov chain]]s); to the theory of [[dynamical systems]] ([[subshifts of finite type]]); to economics ([[Okishio's theorem]],<ref>{{Cite journal|last=Bowles|first=Samuel|date=1981-06-01|title=Technical change and the profit rate: a simple proof of the Okishio theorem|journal=Cambridge Journal of Economics|language=en|volume=5|issue=2|pages=183–186|doi=10.1093/oxfordjournals.cje.a035479|issn=0309-166X}}</ref> [[Hawkins–Simon condition]]<ref name="Meyer681">{{harvnb|Meyer|2000|pp=[http://www.matrixanalysis.com/Chapter8.pdf 8.3.6 p. 681] {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref>); to demography ([[Leslie matrix|Leslie population age distribution model]]);<ref name="Meyer683">{{harvnb|Meyer|2000|pp=[http://www.matrixanalysis.com/Chapter8.pdf 8.3.7 p. 683] {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref> to social networks ([[DeGroot learning|DeGroot learning process]]); to Internet search engines ([[PageRank]]);<ref name="LangvilleMeyer167">{{harvnb|Langville|Meyer|2006|p=[https://books.google.com/books?id=hxvB14-I0twC&pg=PA167 15.2 p. 167]}} {{cite book |url=https://books.google.com/books?id=hxvB14-I0twC&pg=PA167 |title=Google's PageRank and Beyond: The Science of Search Engine Rankings |access-date=2016-10-31 |url-status=bot: unknown |archive-url=https://web.archive.org/web/20140710041730/https://books.google.com/books?id=hxvB14-I0twC&lpg=PP1&dq=isbn%3A0691122024&pg=PA167 |archive-date=July 10, 2014 |isbn=978-0691122021 |last1=Langville |first1=Amy N.|author1-link= Amy Langville |last2=Langville |first2=Amy N. |last3=Meyer |first3=Carl D. |date=2006-07-23 |publisher=Princeton University Press }}</ref> and even to ranking of American football teams.<ref name="Keener80">{{harvnb|Keener|1993|p=[https://www.jstor.org/stable/2132526 p. 80]}}</ref> The first to discuss the ordering of players within tournaments using Perron–Frobenius eigenvectors is [[Edmund Landau]].<ref>{{citation | title = Zur relativen Wertbemessung der Turnierresultaten| pages = 366–369| volume = XI | journal = Deutsches Wochenschach | year=1895 | first1=Edmund | last1=Landau }}</ref><ref>{{citation | title = Über Preisverteilung bei Spielturnieren| pages = 192–202| volume = 63 | journal =Zeitschrift für Mathematik und Physik | year=1915 | first1=Edmund | last1=Landau | url = http://iris.univ-lille1.fr/handle/1908/2031 }}</ref> ==Statement== Let [[nonnegative matrix|'''positive''']] and '''non-negative''' respectively describe [[matrix (mathematics)|matrices]] with exclusively [[positive number|positive]] real numbers as elements and matrices with exclusively non-negative real numbers as elements. The [[eigenvalue]]s of a real [[square matrix]] ''A'' are [[complex numbers]] that make up the [[spectrum of a matrix|spectrum]] of the matrix. The [[exponential growth|exponential growth rate]] of the matrix powers ''A''<sup>''k''</sup> as ''k'' → ∞ is controlled by the eigenvalue of ''A'' with the largest [[absolute value]] ([[Absolute value|modulus]]). The Perron–Frobenius theorem describes the properties of the leading eigenvalue and of the corresponding eigenvectors when ''A'' is a non-negative real square matrix. Early results were due to {{harvs|txt|authorlink=Oskar Perron|first=Oskar|last= Perron|year=1907}} and concerned positive matrices. Later, {{harvs|txt|authorlink=Georg Frobenius|first=Georg |last=Frobenius|year=1912}} found their extension to certain classes of non-negative matrices. ===Positive matrices=== Let <math>A = (a_{ij}) </math> be an <math> n \times n </math> positive matrix: <math> a_{ij} > 0 </math> for <math> 1 \le i,j \le n </math>. Then the following statements hold. # There is a positive real number ''r'', called the '''Perron root''' or the '''Perron–Frobenius eigenvalue''' (also called the '''leading eigenvalue''', '''principal eigenvalue''' or '''dominant eigenvalue'''), such that ''r'' is an eigenvalue of ''A'' and any other eigenvalue ''λ'' (possibly [[complex number|complex]]) in [[absolute value]] is strictly smaller than ''r'' , |''λ''| < ''r''. Thus, the [[spectral radius]] <math>\rho(A) </math> is equal to ''r''. If the matrix coefficients are algebraic, this implies that the eigenvalue is a [[Perron number]]. # The Perron–Frobenius eigenvalue is simple: ''r'' is a simple root of the [[characteristic polynomial]] of ''A''. Consequently, the [[eigenspace]] associated to ''r'' is one-dimensional. (The same is true for the left eigenspace, i.e., the eigenspace for ''A<sup>T</sup>'', the transpose of ''A''.) # There exists an eigenvector ''v'' = (''v''<sub>1</sub>,...,''v''<sub>''n''</sub>)''<sup>T</sup>'' of ''A'' with eigenvalue ''r'' such that all components of ''v'' are positive: ''A v'' = ''r v'', ''v''<sub>''i''</sub> > 0 for 1 ≤ ''i'' ≤ ''n''. (Respectively, there exists a positive left eigenvector ''w'' : ''w<sup>T</sup> A'' = ''w<sup>T</sup>'' r, ''w''<sub>''i''</sub> > 0.) It is known in the literature under many variations as the '''Perron vector''', '''Perron eigenvector''', '''Perron-Frobenius eigenvector''', '''leading eigenvector''', '''principal eigenvector''' or '''dominant eigenvector'''. # There are no other positive (moreover non-negative) eigenvectors except positive multiples of ''v'' (respectively, left eigenvectors except ww'w''), i.e., all other eigenvectors must have at least one negative or non-real component. # <math> \lim_{k \rightarrow \infty} A^k/r^k = v w^T</math>, where the left and right eigenvectors for ''A'' are normalized so that ''w<sup>T</sup>v'' = 1. Moreover, the matrix ''vw<sup>T</sup>'' is the [[Jordan canonical form#Invariant subspace decompositions|projection onto the eigenspace]] corresponding to ''r''. This projection is called the '''Perron projection'''. # '''[[Lothar Collatz|Collatz]]–Wielandt formula''': for all non-negative non-zero vectors ''x'', let ''f''(''x'') be the minimum value of [''Ax'']<sub>''i''</sub> / ''x''<sub>''i''</sub> taken over all those ''i'' such that ''x<sub>i</sub>'' ≠ 0. Then ''f'' is a real valued function whose [[maximum]] over all non-negative non-zero vectors ''x'' is the Perron–Frobenius eigenvalue. # A "Min-max" Collatz–Wielandt formula takes a form similar to the one above: for all strictly positive vectors ''x'', let ''g''(''x'') be the maximum value of [''Ax'']<sub>''i''</sub> / ''x''<sub>''i''</sub> taken over ''i''. Then ''g'' is a real valued function whose [[minimum]] over all strictly positive vectors ''x'' is the Perron–Frobenius eigenvalue. # '''[[Garrett Birkhoff|Birkhoff]]–[[Richard S. Varga|Varga]] formula''': Let ''x'' and ''y'' be strictly positive vectors. Then,<ref>Birkhoff, Garrett and Varga, Richard S., 1958. Reactor criticality and nonnegative matrices. Journal of the Society for Industrial and Applied Mathematics, 6(4), pp.354-377.</ref><math display="block">r = \sup_{x>0} \inf_{y>0} \frac{y^\top A x}{y^\top x} = \inf_{x>0} \sup_{y>0} \frac{y^\top A x}{y^\top x} = \inf_{x>0} \sup_{y>0} \sum_{i,j=1}^n y_i a_{ij} x_j/\sum_{i=1}^n y_i x_i.</math> # '''[[Monroe D. Donsker|Donsker]]–[[S. R. Srinivasa Varadhan|Varadhan]]–[[Shmuel Friedland|Friedland]] formula''': Let ''p'' be a probability vector and ''x'' a strictly positive vector. Then,<ref>Donsker, M.D. and Varadhan, S.S., 1975. On a variational formula for the principal eigenvalue for operators with maximum principle. Proceedings of the National Academy of Sciences, 72(3), pp.780-783.</ref><ref>Friedland, S., 1981. Convex spectral functions. Linear and multilinear algebra, 9(4), pp.299-316.</ref> <math display="block">r = \sup_p \inf_{x>0} \sum_{i=1}^n p_i[Ax]_i/x_i.</math> # '''[[Miroslav Fiedler|Fiedler]] formula''':<ref>{{cite journal |author1=Miroslav Fiedler |author2=Charles R. Johnson |author3=Thomas L. Markham |author4=Michael Neumann |title=A Trace Inequality for M-matrices and the Symmetrizability of a Real Matrix by a Positive Diagonal Matrix |journal=Linear Algebra and Its Applications |date=1985 |volume=71 |pages=81–94 |doi=10.1016/0024-3795(85)90237-X |doi-access=free }}</ref> <math display="block">r = \sup_{z > 0} \ \inf_{x>0, \ y>0,\ x \circ y = z} \frac{y^\top A x}{y^\top x} = \sup_{z > 0} \ \inf_{x>0, \ y>0,\ x \circ y = z}\sum_{i,j=1}^n y_i a_{ij} x_j/\sum_{i=1}^n y_i x_i.</math> # The Perron–Frobenius eigenvalue satisfies the inequalities <math display="block">\min_i \sum_{j} a_{ij} \le r \le \max_i \sum_{j} a_{ij}.</math> All of these properties extend beyond strictly positive matrices to '''primitive matrices''' (see below). Facts 1–7 can be found in Meyer<ref name="Meyer"/> [https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf chapter 8] claims 8.2.11–15 page 667 and exercises 8.2.5,7,9 pages 668–669. The left and right eigenvectors ''w'' and ''v'' are sometimes normalized so that the sum of their components is equal to 1; in this case, they are sometimes called '''stochastic eigenvectors'''. Often they are normalized so that the right eigenvector ''v'' sums to one, while <math>w^T v=1</math>. ===Non-negative matrices=== There is an extension to matrices with non-negative entries. Since any non-negative matrix can be obtained as a limit of positive matrices, one obtains the existence of an eigenvector with non-negative components; the corresponding eigenvalue will be non-negative and greater than ''or equal'', in absolute value, to all other eigenvalues.<ref>{{harvnb|Meyer|2000|pp=[http://www.matrixanalysis.com/Chapter8.pdf chapter 8.3 page 670]. {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref><ref>{{harvnb|Gantmacher|2000|p=[https://books.google.com/books?id=cyX32q8ZP5cC&q=preceding%20section&pg=PA66 chapter XIII.3 theorem 3 page 66]}}</ref> However, for the example <math>A = \left(\begin{smallmatrix}0 & 1\\ 1 & 0\end{smallmatrix}\right)</math>, the maximum eigenvalue ''r'' = 1 has the same absolute value as the other eigenvalue −1; while for <math>A = \left(\begin{smallmatrix}0 & 1\\ 0 & 0\end{smallmatrix}\right)</math>, the maximum eigenvalue is ''r'' = 0, which is not a simple root of the characteristic polynomial, and the corresponding eigenvector (1, 0) is not strictly positive. However, Frobenius found a special subclass of non-negative matrices — ''irreducible'' matrices — for which a non-trivial generalization is possible. For such a matrix, although the eigenvalues attaining the maximal absolute value might not be unique, their structure is under control: they have the form <math>\omega r</math>, where ''<math>r</math>'' is a real strictly positive eigenvalue, and <math>\omega</math> ranges over the complex ''h''' th [[root of unity|roots of 1]] for some positive integer ''h'' called the [[iterated function|period]] of the matrix. The eigenvector corresponding to ''<math>r</math>'' has strictly positive components (in contrast with the general case of non-negative matrices, where components are only non-negative). Also all such eigenvalues are simple roots of the characteristic polynomial. Further properties are described below. ====Classification of matrices==== Let ''A'' be a ''n'' × ''n'' square matrix over [[field (mathematics)|field]] ''F''. The matrix ''A'' is '''irreducible''' if any of the following equivalent properties holds. '''Definition 1 :''' ''A'' does not have non-trivial invariant ''coordinate'' subspaces. Here a non-trivial coordinate subspace means a [[linear subspace]] spanned by any [[proper subset]] of standard basis vectors of ''F<sup>n</sup>''. More explicitly, for any linear subspace spanned by standard basis vectors ''e''<sub>''i''<sub>1</sub> </sub>, ..., ''e''<sub>''i''<sub>k</sub></sub>, 0 < ''k'' < ''n'' its image under the action of ''A'' is not contained in the same subspace. '''Definition 2:''' ''A'' cannot be conjugated into block upper triangular form by a [[permutation matrix]] ''P'': : <math>PAP^{-1} \ne \begin{pmatrix} E & F \\ O & G \end{pmatrix},</math> where ''E'' and ''G'' are non-trivial (i.e. of size greater than zero) square matrices. '''Definition 3:''' One can associate with a matrix ''A'' a certain [[directed graph]] ''G''<sub>''A''</sub>. It has ''n'' vertices labeled 1,...,''n'', and there is an edge from vertex ''i'' to vertex ''j'' precisely when ''a''<sub>''ij''</sub> ≠ 0. Then the matrix ''A'' is irreducible if and only if its associated graph ''G''<sub>''A''</sub> is [[strongly connected component|strongly connected]]. If ''F'' is the field of real or complex numbers, then we also have the following condition. '''Definition 4:''' The [[group representation]] of <math>(\mathbb R, +)</math> on <math>\mathbb{R}^n</math> or <math>(\mathbb C, +)</math> on <math>\mathbb{C}^n</math> given by <math>t \mapsto\exp(tA)</math> has no non-trivial invariant coordinate subspaces. (By comparison, this would be an [[irreducible representation]] if there were no non-trivial invariant subspaces at all, not only considering coordinate subspaces.) A matrix is '''reducible''' if it is not irreducible. A real matrix ''A'' is '''primitive''' if it is non-negative and its ''m''th power is positive for some natural number ''m'' (i.e. all entries of ''A<sup>m</sup>'' are positive). Let ''A'' be real and non-negative. Fix an index ''i'' and define the '''period of index ''i'' ''' to be the [[greatest common divisor]] of all natural numbers ''m'' such that (''A''<sup>''m''</sup>)<sub>''ii''</sub> > 0. When ''A'' is irreducible, the period of every index is the same and is called the '''period of ''A''. ''' In fact, when ''A'' is irreducible, the period can be defined as the greatest common divisor of the lengths of the closed directed paths in ''G''<sub>''A''</sub> (see Kitchens<ref name="Kitchens"/> page 16). The period is also called the index of imprimitivity (Meyer<ref name="Meyer"/> page 674) or the order of cyclicity. If the period is 1, ''A'' is '''aperiodic'''. It can be proved that primitive matrices are the same as irreducible aperiodic non-negative matrices. All statements of the Perron–Frobenius theorem for positive matrices remain true for primitive matrices. The same statements also hold for a non-negative irreducible matrix, except that it may possess several eigenvalues whose absolute value is equal to its spectral radius, so the statements need to be correspondingly modified. In fact the number of such eigenvalues is equal to the period. Results for non-negative matrices were first obtained by Frobenius in 1912. ====Perron–Frobenius theorem for irreducible non-negative matrices==== Let <math>A</math> be an irreducible non-negative <math>N\times N</math> matrix with period <math>h</math> and [[spectral radius]] <math>\rho(A) = r</math>. Then the following statements hold. * The number <math>r\in\mathbb{R}^+</math> is a positive real number and it is an eigenvalue of the matrix <math>A</math>. It is called '''Perron–Frobenius eigenvalue'''. * The Perron–Frobenius eigenvalue <math>r</math> is [[Eigenvalues and eigenvectors#Algebraic multiplicity|simple]]. Both right and left eigenspaces associated with <math>r</math> are one-dimensional. * <math>A</math> has both a right and a left eigenvectors, respectively <math>\mathbf v</math> and <math>\mathbf w</math>, with eigenvalue <math>r</math> and whose components are all positive. Moreover the '''only''' eigenvectors whose components are all positive are those associated with the eigenvalue <math>r</math>. * The matrix <math>A</math> has exactly <math>h</math> (where <math>h</math> is the '''period''') complex eigenvalues with absolute value <math>r</math>. Each of them is a simple root of the characteristic polynomial and is the product of <math>r</math> with an <math>h</math>th [[root of unity]]. * Let <math>\omega = 2\pi/h</math>. Then the matrix <math>A</math> is [[similar matrix|similar]] to <math>e^{i\omega}A</math>, consequently the spectrum of <math>A</math> is invariant under multiplication by <math>e^{i\omega}</math> (i.e. to rotations of the complex plane by the angle <math>\omega</math>). * If <math>h>1</math> then there exists a permutation matrix <math>P</math> such that ::<math>PAP^{-1}= \begin{pmatrix} O & A_1 & O & O & \ldots & O \\ O & O & A_2 & O & \ldots & O \\ \vdots & \vdots &\vdots & \vdots & & \vdots \\ O & O & O & O & \ldots & A_{h-1} \\ A_h & O & O & O & \ldots & O \end{pmatrix}, </math> :: where <math>O</math> denotes a zero matrix and the blocks along the main diagonal are square matrices. * '''[[Lothar Collatz|Collatz]]–Wielandt formula''': for all non-negative non-zero vectors ''<math>\mathbf x </math>'' let ''<math>f(\mathbf x) </math>'' be the minimum value of ''<math>[A\mathbf x]_i/x_i </math>'' taken over all those <math>i </math> such that <math>x_i\neq0 </math>. Then <math>f </math> is a real valued function whose [[maximum]] is the Perron–Frobenius eigenvalue. * The Perron–Frobenius eigenvalue satisfies the inequalities ::<math>\min_i \sum_{j} a_{ij} \le r \le \max_i \sum_{j} a_{ij}.</math> The example <math>A =\left(\begin{smallmatrix} 0 & 0 & 1 \\ 0 & 0 & 1 \\ 1 & 1 & 0 \end{smallmatrix}\right)</math> shows that the (square) zero-matrices along the diagonal may be of different sizes, the blocks ''A''<sub>''j''</sub> need not be square, and ''h'' need not divide ''n''. ===Further properties=== Let ''A'' be an irreducible non-negative matrix, then: # (I+''A'')<sup>''n''−1</sup> is a positive matrix. (Meyer<ref name="Meyer"/> [https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf claim 8.3.5 p. 672]). For a non-negative ''A'', this is also a sufficient condition.<ref name="Minc">{{cite book |first=Henryk |last= Minc |author-link=Henryk Minc |title=Nonnegative matrices |isbn=0-471-83966-3 |year=1988 |publisher=John Wiley & Sons |location=New York |page=6 [Corollary 2.2] }}</ref> # Wielandt's theorem.<ref>{{Cite book |author=Gradshtein, Izrailʹ Solomonovich |url=http://worldcat.org/oclc/922964628 |title=Table of integrals, series, and products |date=18 September 2014 |publisher=Elsevier |isbn=978-0-12-384934-2 |oclc=922964628}}</ref>{{clarify|reason=What are the restrictions on ''B''?|date=March 2015}} If |''B''|<''A'', then ''ρ''(''B'')≤''ρ''(''A''). If equality holds (i.e. if ''μ=ρ(A)e<sup>iφ</sup>'' is eigenvalue for ''B''), then ''B'' = ''e''<sup>''iφ''</sup> ''D AD''<sup>−1</sup> for some diagonal unitary matrix ''D'' (i.e. diagonal elements of ''D'' equals to ''e''<sup>''iΘ''<sub>''l''</sub></sup>, non-diagonal are zero).<ref name="Meyer675">{{harvnb|Meyer|2000|pp=[http://www.matrixanalysis.com/Chapter8.pdf claim 8.3.11 p. 675] {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref> # If some power ''A<sup>q</sup>'' is reducible, then it is completely reducible, i.e. for some permutation matrix ''P'', it is true that: <math> P A^q P^{-1}= \begin{pmatrix} A_1 & O & O & \dots & O \\ O & A_2 & O & \dots & O \\ \vdots & \vdots & \vdots & & \vdots \\ O & O & O & \dots & A_d \\ \end{pmatrix} </math>, where ''A<sub>i</sub>'' are irreducible matrices having the same maximal eigenvalue. The number of these matrices ''d'' is the greatest common divisor of ''q'' and ''h'', where ''h'' is period of ''A''.<ref>{{harvnb|Gantmacher|2000|p=section XIII.5 theorem 9}}</ref> # If ''c''(''x'') ''= x<sup>n</sup> + c<sub>k<sub>1</sub></sub> x<sup>n-k<sub>1</sub></sup> + c<sub>k<sub>2</sub></sub> x<sup>n-k<sub>2</sub></sup> + ... + c<sub>k<sub>s</sub></sub> x<sup>n-k<sub>s</sub></sup>'' is the characteristic polynomial of ''A'' in which only the non-zero terms are listed, then the period of ''A'' equals the greatest common divisor of ''k<sub>1</sub>, k<sub>2</sub>, ... , k<sub>s</sub>''.<ref>{{harvnb|Meyer|2000|pp=[http://www.matrixanalysis.com/Chapter8.pdf page 679] {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref> # [[Cesàro summation|Cesàro]] [[summability theory|averages]]: <math> \lim_{k \rightarrow \infty} 1/k\sum_{i=0,...,k} A^i/r^i = ( v w^T),</math> where the left and right eigenvectors for ''A'' are normalized so that ''w''<sup>''T''</sup>''v'' = 1. Moreover, the matrix ''v w<sup>T</sup>'' is the [[Spectral theorem|spectral projection]] corresponding to ''r'', the Perron projection.<ref>{{harvnb|Meyer|2000|pp=[http://www.matrixanalysis.com/Chapter8.pdf example 8.3.2 p. 677] {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref> # Let ''r'' be the Perron–Frobenius eigenvalue, then the adjoint matrix for (''r''-''A'') is positive.<ref>{{harvnb|Gantmacher|2000|p=[https://books.google.com/books?id=cyX32q8ZP5cC&q=preceding%20section&pg=PA62 section XIII.2.2 page 62]}}</ref> # If ''A'' has at least one non-zero diagonal element, then ''A'' is primitive.<ref>{{harvnb|Meyer|2000|pp= [http://www.matrixanalysis.com/Chapter8.pdf example 8.3.3 p. 678] {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref> # If 0 ≤ ''A'' < ''B'', then ''r''<sub>''A''</sub> ≤ ''r''<sub>''B.''</sub> Moreover, if ''B'' is irreducible, then the inequality is strict: ''r<sub>A</sub> < r<sub>B</sub>''. A matrix ''A'' is primitive provided it is non-negative and ''A<sup>m</sup>'' is positive for some ''m'', and hence ''A<sup>k</sup>'' is positive for all ''k ≥ m''. To check primitivity, one needs a bound on how large the minimal such ''m'' can be, depending on the size of ''A'':<ref>{{harvnb|Meyer|2000|pp=[http://www.matrixanalysis.com/Chapter8.pdf chapter 8 example 8.3.4 page 679 and exercise 8.3.9 p. 685] {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref> * If ''A'' is a non-negative primitive matrix of size ''n'', then ''A''<sup>''n''<sup>2</sup> − 2''n'' + 2</sup> is positive. Moreover, this is the best possible result, since for the matrix ''M'' below, the power ''M<sup>k</sup>'' is not positive for every ''k'' < ''n''<sup>2</sup> − 2''n'' + 2, since (''M''<sup>''n''<sup>2</sup> − 2''n''+1</sup>)<sub>1,1</sub> = 0. :<math>M= \left(\begin{smallmatrix} 0 & 1 & 0 & 0 & \cdots & 0 \\ 0 & 0 & 1 & 0 & \cdots & 0 \\ 0 & 0 & 0 & 1 & \cdots & 0 \\ \vdots & \vdots & \vdots & \vdots & & \vdots \\ 0 & 0 & 0 & 0 & \cdots & 1 \\ 1 & 1 & 0 & 0 & \cdots & 0 \end{smallmatrix}\right) </math> ==Applications== Numerous books have been written on the subject of non-negative matrices, and Perron–Frobenius theory is invariably a central feature. The following examples given below only scratch the surface of its vast application domain. ===Non-negative matrices=== The Perron–Frobenius theorem does not apply directly to non-negative matrices. Nevertheless, any reducible square matrix ''A'' may be written in upper-triangular block form (known as the '''normal form of a reducible matrix''')<ref>{{harvnb|Varga|2002|p=2.43 (page 51)}}</ref> ::::''PAP''<sup>−1</sup> = <math> \left( \begin{smallmatrix} B_1 & * & * & \cdots & * \\ 0 & B_2 & * & \cdots & * \\ \vdots & \vdots & \vdots & & \vdots \\ 0 & 0 & 0 & \cdots & * \\ 0 & 0 & 0 & \cdots & B_h \end{smallmatrix} \right)</math> where ''P'' is a permutation matrix and each ''B<sub>i</sub>'' is a square matrix that is either irreducible or zero. Now if ''A'' is non-negative then so too is each block of ''PAP''<sup>−1</sup>, moreover the spectrum of ''A'' is just the union of the spectra of the ''B<sub>i</sub>''. The invertibility of ''A'' can also be studied. The inverse of ''PAP''<sup>−1</sup> (if it exists) must have diagonal blocks of the form ''B<sub>i</sub>''<sup>−1</sup> so if any ''B<sub>i</sub>'' isn't invertible then neither is ''PAP''<sup>−1</sup> or ''A''. Conversely let ''D'' be the block-diagonal matrix corresponding to ''PAP''<sup>−1</sup>, in other words ''PAP''<sup>−1</sup> with the asterisks zeroised. If each ''B<sub>i</sub>'' is invertible then so is ''D'' and ''D''<sup>−1</sup>(''PAP''<sup>−1</sup>) is equal to the identity plus a nilpotent matrix. But such a matrix is always invertible (if ''N<sup>k</sup>'' = 0 the inverse of 1 − ''N'' is 1 + ''N'' + ''N''<sup>2</sup> + ... + ''N''<sup>''k''−1</sup>) so ''PAP''<sup>−1</sup> and ''A'' are both invertible. Therefore, many of the spectral properties of ''A'' may be deduced by applying the theorem to the irreducible ''B<sub>i</sub>''. For example, the Perron root is the maximum of the ρ(''B<sub>i</sub>''). While there will still be eigenvectors with non-negative components it is quite possible that none of these will be positive. ===Stochastic matrices=== A row (column) [[stochastic matrix]] is a square matrix each of whose rows (columns) consists of non-negative real numbers whose sum is unity. The theorem cannot be applied directly to such matrices because they need not be irreducible. If ''A'' is row-stochastic then the column vector with each entry 1 is an eigenvector corresponding to the eigenvalue 1, which is also ρ(''A'') by the remark above. It might not be the only eigenvalue on the unit circle: and the associated eigenspace can be multi-dimensional. If ''A'' is row-stochastic and irreducible then the Perron projection is also row-stochastic and all its rows are equal. ===Algebraic graph theory=== The theorem has particular use in [[algebraic graph theory]]. The "underlying graph" of a nonnegative ''n''-square matrix is the graph with vertices numbered 1, ..., ''n'' and arc ''ij'' if and only if ''A<sub>ij</sub>'' ≠ 0. If the underlying graph of such a matrix is strongly connected, then the matrix is irreducible, and thus the theorem applies. In particular, the [[adjacency matrix]] of a [[strongly connected component|strongly connected graph]] is irreducible.<ref>{{cite book |author-link=Richard A. Brualdi |first1=Richard A. |last1=Brualdi |author-link2=H. J. Ryser |first2=Herbert J. |last2=Ryser |title=Combinatorial Matrix Theory |url=https://archive.org/details/combinatorialmat0000brua_x9u3 |url-access=registration |location=Cambridge |publisher=Cambridge UP |year=1992 |isbn=978-0-521-32265-2 }}</ref><ref>{{cite book |author-link=Richard A. Brualdi |first1=Richard A. |last1=Brualdi |first2=Dragos |last2=Cvetkovic |title=A Combinatorial Approach to Matrix Theory and Its Applications |publisher=CRC Press |location=Boca Raton, FL |year=2009 |isbn=978-1-4200-8223-4 }}</ref> ===Finite Markov chains=== The theorem has a natural interpretation in the theory of finite [[Markov chain]]s (where it is the matrix-theoretic equivalent of the convergence of an irreducible finite Markov chain to its stationary distribution, formulated in terms of the transition matrix of the chain; see, for example, the article on the [[subshift of finite type]]).<!--which article?--> ===Compact operators=== {{main|Krein–Rutman theorem}} More generally, it can be extended to the case of non-negative [[compact operator]]s, which, in many ways, resemble finite-dimensional matrices. These are commonly studied in physics, under the name of [[transfer operator]]s, or sometimes '''Ruelle–Perron–Frobenius operators''' (after [[David Ruelle]]). In this case, the leading eigenvalue corresponds to the [[thermodynamic equilibrium]] of a [[dynamical system]], and the lesser eigenvalues to the decay modes of a system that is not in equilibrium. Thus, the theory offers a way of discovering the [[arrow of time]] in what would otherwise appear to be reversible, deterministic dynamical processes, when examined from the point of view of [[point-set topology]].<ref>{{cite book |first=Michael C. |last=Mackey |title=Time's Arrow: The origins of thermodynamic behaviour |location=New York |publisher=Springer-Verlag |year=1992 |isbn=978-0-387-97702-7 }}</ref> ==Proof methods== A common thread in many proofs is the [[Brouwer fixed point theorem]]. Another popular method is that of Wielandt (1950). He used the [[Lothar Collatz|Collatz]]–Wielandt formula described above to extend and clarify Frobenius's work.<ref>{{harvnb|Gantmacher|2000|p=[https://books.google.com/books?id=cyX32q8ZP5cC&dq=Applications%20of%20the%20theory%20of%20matrices&pg=PA54 section XIII.2.2 page 54]}}</ref> Another proof is based on the [[spectral theory]]<ref name="Smith"/> from which part of the arguments are borrowed. ===Perron root is strictly maximal eigenvalue for positive (and primitive) matrices=== If ''A'' is a positive (or more generally primitive) matrix, then there exists a real positive eigenvalue ''r'' (Perron–Frobenius eigenvalue or Perron root), which is strictly greater in absolute value than all other eigenvalues, hence ''r'' is the [[spectral radius]] of ''A''. This statement does not hold for general non-negative irreducible matrices, which have ''h'' eigenvalues with the same absolute eigenvalue as ''r'', where ''h'' is the period of ''A''. ====Proof for positive matrices==== Let ''A'' be a positive matrix, assume that its spectral radius ρ(''A'') = 1 (otherwise consider ''A/ρ(A)''). Hence, there exists an eigenvalue λ on the unit circle, and all the other eigenvalues are less or equal 1 in absolute value. Suppose that another eigenvalue λ ≠ 1 also falls on the unit circle. Then there exists a positive integer ''m'' such that ''A<sup>m</sup>'' is a positive matrix and the real part of λ''<sup>m</sup>'' is negative. Let ε be half the smallest diagonal entry of ''A<sup>m</sup>'' and set ''T'' = ''A<sup>m</sup>'' − ''εI'' which is yet another positive matrix. Moreover, if ''Ax'' = ''λx'' then ''A<sup>m</sup>x'' = ''λ<sup>m</sup>x'' thus ''λ''<sup>''m''</sup> − ''ε'' is an eigenvalue of ''T''. Because of the choice of ''m'' this point lies outside the unit disk consequently ''ρ''(''T'') > 1. On the other hand, all the entries in ''T'' are positive and less than or equal to those in ''A<sup>m</sup>'' so by [[spectral radius|Gelfand's formula]] ''ρ''(''T'') ≤ ''ρ''(''A<sup>m</sup>'') ≤ ''ρ''(''A'')<sup>''m''</sup> = 1. This contradiction means that λ=1 and there can be no other eigenvalues on the unit circle. Absolutely the same arguments can be applied to the case of primitive matrices; we just need to mention the following simple lemma, which clarifies the properties of primitive matrices. ====Lemma==== Given a non-negative ''A'', assume there exists ''m'', such that ''A<sup>m</sup>'' is positive, then ''A''<sup>''m''+1</sup>, ''A''<sup>''m''+2</sup>, ''A''<sup>''m''+3</sup>,... are all positive. ''A''<sup>''m''+1</sup> = ''AA''<sup>''m''</sup>, so it can have zero element only if some row of ''A'' is entirely zero, but in this case the same row of ''A<sup>m</sup>'' will be zero. Applying the same arguments as above for primitive matrices, prove the main claim. ===Power method and the positive eigenpair=== For a positive (or more generally irreducible non-negative) matrix ''A'' the dominant [[eigenvector]] is real and strictly positive (for non-negative ''A'' respectively non-negative.) This can be established using the [[power method]], which states that for a sufficiently generic (in the sense below) matrix ''A'' the sequence of vectors ''b''<sub>''k''+1</sub> = ''Ab''<sub>''k''</sub> / | ''Ab''<sub>''k''</sub> | converges to the [[eigenvector]] with the maximum [[eigenvalue]]. (The initial vector ''b''<sub>0</sub> can be chosen arbitrarily except for some measure zero set). Starting with a non-negative vector ''b''<sub>0</sub> produces the sequence of non-negative vectors ''b<sub>k</sub>''. Hence the limiting vector is also non-negative. By the power method this limiting vector is the dominant eigenvector for ''A'', proving the assertion. The corresponding eigenvalue is non-negative. The proof requires two additional arguments. First, the power method converges for matrices which do not have several eigenvalues of the same absolute value as the maximal one. The previous section's argument guarantees this. Second, to ensure strict positivity of all of the components of the eigenvector for the case of irreducible matrices. This follows from the following fact, which is of independent interest: :Lemma: given a positive (or more generally irreducible non-negative) matrix ''A'' and ''v'' as any non-negative eigenvector for ''A'', then it<!--the eigenvector?--> is necessarily strictly positive and the corresponding eigenvalue is also strictly positive. Proof. One of the definitions of irreducibility for non-negative matrices is that for all indexes ''i,j'' there exists ''m'', such that (''A''<sup>''m''</sup>)<sub>''ij''</sub> is strictly positive. Given a non-negative eigenvector ''v'', and that at least one of its components say ''i''-th is strictly positive, the corresponding eigenvalue is strictly positive, indeed, given ''n'' such that (''A''<sup>''n''</sup>)<sub>''ii''</sub> >0, hence: ''r''<sup>''n''</sup>''v''<sub>''i''</sub> = ''A''<sup>''n''</sup>''v''<sub>''i''</sub> ≥ (''A''<sup>''n''</sup>)<sub>''ii''</sub>''v''<sub>''i''</sub> >0. Hence ''r'' is strictly positive. The eigenvector is strict positivity. Then given ''m'', such that (''A''<sup>''m''</sup>)<sub>''ji''</sub> >0, hence: ''r''<sup>''m''</sup>''v''<sub>''j''</sub> = (''A''<sup>''m''</sup>''v'')<sub>''j''</sub> ≥ (''A''<sup>''m''</sup>)<sub>''ji''</sub>''v''<sub>''i''</sub> >0, hence ''v''<sub>''j''</sub> is strictly positive, i.e., the eigenvector is strictly positive. ===Multiplicity one=== This section proves that the Perron–Frobenius eigenvalue is a simple root of the characteristic polynomial of the matrix. Hence the eigenspace associated to Perron–Frobenius eigenvalue ''r'' is one-dimensional. The arguments here are close to those in Meyer.<ref name="Meyer">{{harvnb|Meyer|2000|pp=[http://www.matrixanalysis.com/Chapter8.pdf chapter 8 page 665] {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref> Given a strictly positive eigenvector ''v'' corresponding to ''r'' and another eigenvector ''w'' with the same eigenvalue. (The vectors ''v'' and ''w'' can be chosen to be real, because ''A'' and ''r'' are both real, so the null space of ''A-r'' has a basis consisting of real vectors.) Assuming at least one of the components of ''w'' is positive (otherwise multiply ''w'' by −1). Given maximal possible ''α'' such that ''u=v- α w'' is non-negative, then one of the components of ''u'' is zero, otherwise ''α'' is not maximum. Vector ''u'' is an eigenvector. It is non-negative, hence by the lemma described in the [[#Power method and the positive eigenpair|previous section]] non-negativity implies strict positivity for any eigenvector. On the other hand, as above at least one component of ''u'' is zero. The contradiction implies that ''w'' does not exist. Case: There are no Jordan blocks corresponding to the Perron–Frobenius eigenvalue ''r'' and all other eigenvalues which have the same absolute value. If there is a Jordan block, then the [[Matrix norm#Induced norm|infinity norm]] (A/r)<sup>k</sup><sub>∞</sub> tends to infinity for ''k → ∞ '', but that contradicts the existence of the positive eigenvector. Given ''r'' = 1, or ''A/r''. Letting ''v'' be a Perron–Frobenius strictly positive eigenvector, so ''Av=v'', then: <math> \|v\|_{\infty}= \|A^k v\|_{\infty} \ge \|A^k\|_{\infty} \min_i (v_i), ~~\Rightarrow~~ \|A^k\|_{\infty} \le \|v\|/\min_i (v_i) </math> So ‖''A<sup>k</sup>''‖<sub>∞</sub> is bounded for all ''k''. This gives another proof that there are no eigenvalues which have greater absolute value than Perron–Frobenius one. It also contradicts the existence of the Jordan block for any eigenvalue which has absolute value equal to 1 (in particular for the Perron–Frobenius one), because existence of the Jordan block implies that ‖''A<sup>k</sup>''‖<sub>∞</sub> is unbounded. For a two by two matrix: : <math> J^k= \begin{pmatrix} \lambda & 1 \\ 0 & \lambda \end{pmatrix} ^k = \begin{pmatrix} \lambda^k & k\lambda^{k-1} \\ 0 & \lambda^k \end{pmatrix}, </math> hence ‖''J''<sup>''k''</sup>‖<sub>∞</sub> = |''k'' + ''λ''| (for |''λ''| = 1), so it tends to infinity when ''k'' does so. Since ''J<sup>k</sup>'' = ''C''<sup>−1</sup> ''A''<sup>''k''</sup>''C'', then ''A''<sup>''k''</sup> ≥ ''J''<sup>''k''</sup>/ (''C''<sup>−1</sup> ''C'' ), so it also tends to infinity. The resulting contradiction implies that there are no Jordan blocks for the corresponding eigenvalues. Combining the two claims above reveals that the Perron–Frobenius eigenvalue ''r'' is simple root of the characteristic polynomial. In the case of nonprimitive matrices, there exist other eigenvalues which have the same absolute value as ''r''. The same claim is true for them, but requires more work. ===No other non-negative eigenvectors=== Given positive (or more generally irreducible non-negative matrix) ''A'', the Perron–Frobenius eigenvector is the only (up to multiplication by constant) non-negative eigenvector for ''A''. Other eigenvectors must contain negative or complex components since eigenvectors for different eigenvalues are orthogonal in some sense, but two positive eigenvectors cannot be orthogonal, so they must correspond to the same eigenvalue, but the eigenspace for the Perron–Frobenius is one-dimensional. Assuming there exists an eigenpair (''λ'', ''y'') for ''A'', such that vector ''y'' is positive, and given (''r'', ''x''), where ''x'' – is the left Perron–Frobenius eigenvector for ''A'' (i.e. eigenvector for ''A<sup>T</sup>''), then ''rx''<sup>''T''</sup>''y'' = (''x''<sup>''T''</sup> ''A'') ''y'' = ''x''<sup>''T''</sup> (''Ay'') = ''λx''<sup>''T''</sup>''y'', also ''x''<sup>''T''</sup> ''y'' > 0, so one has: ''r'' = ''λ''. Since the eigenspace for the Perron–Frobenius eigenvalue ''r'' is one-dimensional, non-negative eigenvector ''y'' is a multiple of the Perron–Frobenius one.<ref>{{harvnb|Meyer|2000|pp=[http://www.matrixanalysis.com/Chapter8.pdf chapter 8 claim 8.2.10 page 666] {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref> ===Collatz–Wielandt formula=== Given a positive (or more generally irreducible non-negative matrix) ''A'', one defines the function ''f'' on the set of all non-negative non-zero vectors ''x'' such that ''f(x)'' is the minimum value of [''Ax'']<sub>''i''</sub> / ''x''<sub>''i''</sub> taken over all those ''i'' such that ''x<sub>i</sub>'' ≠ 0. Then ''f'' is a real-valued function, whose [[maximum]] is the Perron–Frobenius eigenvalue ''r''. For the proof we denote the maximum of ''f'' by the value ''R''. The proof requires to show '' R = r''. Inserting the Perron-Frobenius eigenvector ''v'' into ''f'', we obtain ''f(v) = r'' and conclude ''r ≤ R''. For the opposite inequality, we consider an arbitrary nonnegative vector ''x'' and let ''ξ=f(x)''. The definition of ''f'' gives ''0 ≤ ξx ≤ Ax'' (componentwise). Now, we use the positive right eigenvector ''w'' for ''A'' for the Perron-Frobenius eigenvalue ''r'', then '' ξ w<sup>T</sup> x = w<sup>T</sup> ξx ≤ w<sup>T</sup> (Ax) = (w<sup>T</sup> A)x = r w<sup>T</sup> x ''. Hence ''f(x) = ξ ≤ r'', which implies ''R ≤ r''.<ref>{{harvnb|Meyer|2000|pp=[http://www.matrixanalysis.com/Chapter8.pdf chapter 8 page 666] {{cite web |url=http://www.matrixanalysis.com/Chapter8.pdf |title=Archived copy |access-date=2010-03-07 |url-status=dead |archive-url=https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date=March 7, 2010 }}}}</ref> ===Perron projection as a limit: ''A''<sup>''k''</sup>/''r''<sup>''k''</sup>=== Let ''A'' be a positive (or more generally, primitive) matrix, and let ''r'' be its Perron–Frobenius eigenvalue. # There exists a limit ''A<sup>k</sup>/r<sup>k</sup>'' for ''k → ∞'', denote it by ''P''. # ''P'' is a [[Projection (linear algebra)|projection operator]]: ''P''<sup>2</sup> = ''P'', which commutes with ''A'': ''AP'' = ''PA''. # The image of ''P'' is one-dimensional and spanned by the Perron–Frobenius eigenvector ''v'' (respectively for ''P<sup>T</sup>''—by the Perron–Frobenius eigenvector ''w'' for ''A<sup>T</sup>''). # ''P'' = ''vw''<sup>''T''</sup>, where ''v,w'' are normalized such that ''w''<sup>''T''</sup> ''v'' = 1. # Hence ''P'' is a positive operator. Hence ''P'' is a [[spectral projection]] for the Perron–Frobenius eigenvalue ''r'', and is called the Perron projection. The above assertion is not true for general non-negative irreducible matrices. Actually the claims above (except claim 5) are valid for any matrix ''M'' such that there exists an eigenvalue ''r'' which is strictly greater than the other eigenvalues in absolute value and is the simple root of the characteristic [[polynomial]]. (These requirements hold for primitive matrices as above). Given that ''M'' is diagonalizable, ''M'' is conjugate to a diagonal matrix with eigenvalues ''r''<sub>1</sub>, ... , ''r''<sub>''n''</sub> on the diagonal (denote ''r''<sub>1</sub> = ''r''). The matrix ''M''<sup>''k''</sup>/''r''<sup>''k''</sup> will be conjugate (1, (''r''<sub>2</sub>/''r'')<sup>''k''</sup>, ... , (''r''<sub>''n''</sub>/''r'')<sup>''k''</sup>), which tends to (1,0,0,...,0), for ''k → ∞'', so the limit exists. The same method works for general ''M'' (without assuming that ''M'' is diagonalizable). The projection and commutativity properties are elementary corollaries of the definition: ''MM''<sup>''k''</sup>/''r''<sup>''k''</sup> = ''M''<sup>''k''</sup>/''r''<sup>''k''</sup> ''M'' ; ''P''<sup>2</sup> = lim ''M''<sup>2''k''</sup>/''r''<sup>2''k''</sup> = ''P''. The third fact is also elementary: ''M''(''Pu'') = ''M'' lim ''M''<sup>''k''</sup>/''r''<sup>''k''</sup> ''u'' = lim ''rM''<sup>''k''+1</sup>/''r''<sup>''k''+1</sup>''u'', so taking the limit yields ''M''(''Pu'') = ''r''(''Pu''), so image of ''P'' lies in the ''r''-eigenspace for ''M'', which is one-dimensional by the assumptions. Denoting by ''v'', ''r''-eigenvector for ''M'' (by ''w'' for ''M<sup>T</sup>''). Columns of ''P'' are multiples of ''v'', because the image of ''P'' is spanned by it. Respectively, rows of ''w''. So ''P'' takes a form ''(a v w<sup>T</sup>)'', for some ''a''. Hence its trace equals to ''(a w<sup>T</sup> v)''. Trace of projector equals the dimension of its image. It was proved before that it is not more than one-dimensional. From the definition one sees that ''P'' acts identically on the ''r''-eigenvector for ''M''. So it is one-dimensional. So choosing (''w''<sup>''T''</sup>''v'') = 1, implies ''P'' = ''vw''<sup>''T''</sup>. ===Inequalities for Perron–Frobenius eigenvalue=== For any non-negative matrix ''A'' its Perron–Frobenius eigenvalue ''r'' satisfies the inequality: :<math> r \; \le \; \max_i \sum_j a_{ij}.</math> This is not specific to non-negative matrices: for any matrix ''A'' with an eigenvalue <math>\scriptstyle\lambda</math> it is true that <math>\scriptstyle |\lambda| \; \le \; \max_i \sum_j |a_{ij}|</math>. This is an immediate corollary of the [[Gershgorin circle theorem]]. However another proof is more direct: Any [[Matrix norm#Induced norm|matrix induced norm]] satisfies the inequality <math>\scriptstyle\|A\| \ge |\lambda|</math> for any eigenvalue <math>\scriptstyle\lambda</math> because, if <math>\scriptstyle x</math> is a corresponding eigenvector, <math>\scriptstyle\|A\| \ge |Ax|/|x| = |\lambda x|/|x| = |\lambda|</math>. The [[Matrix norm#Induced norm|infinity norm]] of a matrix is the maximum of row sums: <math>\scriptstyle \left \| A \right \| _\infty = \max \limits _{1 \leq i \leq m} \sum _{j=1} ^n | a_{ij} |. </math> Hence the desired inequality is exactly <math>\scriptstyle\|A\|_\infty \ge |\lambda|</math> applied to the non-negative matrix ''A''. Another inequality is: :<math>\min_i \sum_j a_{ij} \; \le \; r .</math> This fact is specific to non-negative matrices; for general matrices there is nothing similar. Given that ''A'' is positive (not just non-negative), then there exists a positive eigenvector ''w'' such that ''Aw'' = ''rw'' and the smallest component of ''w'' (say ''w<sub>i</sub>'') is 1. Then ''r'' = (''Aw'')<sub>''i''</sub> ≥ the sum of the numbers in row ''i'' of ''A''. Thus the minimum row sum gives a lower bound for ''r'' and this observation can be extended to all non-negative matrices by continuity. Another way to argue it is via the [[Lothar Collatz|Collatz]]-Wielandt formula. One takes the vector ''x'' = (1, 1, ..., 1) and immediately obtains the inequality. ===Further proofs=== ====Perron projection==== The proof now proceeds using [[Eigendecomposition of a matrix|spectral decomposition]]. The trick here is to split the Perron root from the other eigenvalues. The spectral projection associated with the Perron root is called the Perron projection and it enjoys the following property: The Perron projection of an irreducible non-negative square matrix is a positive matrix. Perron's findings and also (1)–(5) of the theorem are corollaries of this result. The key point is that a positive projection always has rank one. This means that if ''A'' is an irreducible non-negative square matrix then the algebraic and geometric multiplicities of its Perron root are both one. Also if ''P'' is its Perron projection then ''AP'' = ''PA'' = ρ(''A'')''P'' so every column of ''P'' is a positive right eigenvector of ''A'' and every row is a positive left eigenvector. Moreover, if ''Ax'' = λ''x'' then ''PAx'' = λ''Px'' = ρ(''A'')''Px'' which means ''Px'' = 0 if λ ≠ ρ(''A''). Thus the only positive eigenvectors are those associated with ρ(''A''). If ''A'' is a primitive matrix with ρ(''A'') = 1 then it can be decomposed as ''P'' ⊕ (1 − ''P'')''A'' so that ''A<sup>n</sup>'' = ''P'' + (1 − ''P'')''A''<sup>''n''</sup>. As ''n'' increases the second of these terms decays to zero leaving ''P'' as the limit of ''A<sup>n</sup>'' as ''n'' → ∞. The power method is a convenient way to compute the Perron projection of a primitive matrix. If ''v'' and ''w'' are the positive row and column vectors that it generates then the Perron projection is just ''wv''/''vw''. The spectral projections aren't neatly blocked as in the Jordan form. Here they are overlaid and each generally has complex entries extending to all four corners of the square matrix. Nevertheless, they retain their mutual orthogonality which is what facilitates the decomposition. ====Peripheral projection==== The analysis when ''A'' is irreducible and non-negative is broadly similar. The Perron projection is still positive but there may now be other eigenvalues of modulus ρ(''A'') that negate use of the power method and prevent the powers of (1 − ''P'')''A'' decaying as in the primitive case whenever ρ(''A'') = 1. So we consider the '''peripheral projection''', which is the spectral projection of ''A'' corresponding to all the eigenvalues that have modulus ''ρ''(''A''). It may then be shown that the peripheral projection of an irreducible non-negative square matrix is a non-negative matrix with a positive diagonal. ====Cyclicity==== Suppose in addition that ρ(''A'') = 1 and ''A'' has ''h'' eigenvalues on the unit circle. If ''P'' is the peripheral projection then the matrix ''R'' = ''AP'' = ''PA'' is non-negative and irreducible, ''R<sup>h</sup>'' = ''P'', and the cyclic group ''P'', ''R'', ''R''<sup>2</sup>, ...., ''R''<sup>''h''−1</sup> represents the harmonics of ''A''. The spectral projection of ''A'' at the eigenvalue λ on the unit circle is given by the formula <math>\scriptstyle h^{-1}\sum^h_1\lambda^{-k}R^k</math>. All of these projections (including the Perron projection) have the same positive diagonal, moreover choosing any one of them and then taking the modulus of every entry invariably yields the Perron projection. Some donkey work is still needed in order to establish the cyclic properties (6)–(8) but it's essentially just a matter of turning the handle. The spectral decomposition of ''A'' is given by ''A'' = ''R'' ⊕ (1 − ''P'')''A'' so the difference between ''A<sup>n</sup>'' and ''R<sup>n</sup>'' is ''A<sup>n</sup>'' − ''R<sup>n</sup>'' = (1 − ''P'')''A''<sup>''n''</sup> representing the transients of ''A<sup>n</sup>'' which eventually decay to zero. ''P'' may be computed as the limit of ''A<sup>nh</sup>'' as ''n'' → ∞. ==Counterexamples== The matrices ''L'' = <math>\left( \begin{smallmatrix} 1 & 0 & 0 \\ 1 & 0 & 0 \\ 1 & 1 & 1 \end{smallmatrix} \right)</math>, ''P'' = <math>\left( \begin{smallmatrix} 1 & 0 & 0 \\ 1 & 0 & 0 \\ \!-1 & 1 & 1 \end{smallmatrix} \right)</math>, ''T'' = <math>\left( \begin{smallmatrix} 0 & 1 & 1 \\ 1 & 0 & 1 \\ 1 & 1 & 0 \end{smallmatrix} \right)</math>, ''M'' = <math>\left( \begin{smallmatrix} 0 & 1 & 0 & 0 & 0 \\ 1 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 1 & 0 \\ 0 & 0 & 0 & 0 & 1 \\ 0 & 0 & 1 & 0 & 0 \end{smallmatrix} \right)</math> provide simple examples of what can go wrong if the necessary conditions are not met. It is easily seen that the Perron and peripheral projections of ''L'' are both equal to ''P'', thus when the original matrix is reducible the projections may lose non-negativity and there is no chance of expressing them as limits of its powers. The matrix ''T'' is an example of a primitive matrix with zero diagonal. If the diagonal of an irreducible non-negative square matrix is non-zero then the matrix must be primitive but this example demonstrates that the converse is false. ''M'' is an example of a matrix with several missing spectral teeth. If ω = e<sup>iπ/3</sup> then ω<sup>6</sup> = 1 and the eigenvalues of ''M'' are {1,ω<sup>2</sup>,ω<sup>3</sup>=-1,ω<sup>4</sup>} with a dimension 2 eigenspace for +1 so ω and ω<sup>5</sup> are both absent. More precisely, since ''M'' is block-diagonal cyclic, then the eigenvalues are {1,-1} for the first block, and {1,ω<sup>2</sup>,ω<sup>4</sup>} for the lower one{{Citation needed|date=January 2012}} ==Terminology== A problem that causes confusion is a lack of standardisation in the definitions. For example, some authors use the terms ''strictly positive'' and ''positive'' to mean > 0 and ≥ 0 respectively. In this article ''positive'' means > 0 and ''non-negative'' means ≥ 0. Another vexed area concerns ''decomposability'' and ''reducibility'': ''irreducible'' is an overloaded term. For avoidance of doubt a non-zero non-negative square matrix ''A'' such that 1 + ''A'' is primitive is sometimes said to be ''connected''. Then irreducible non-negative square matrices and connected matrices are synonymous.<ref>For surveys of results on irreducibility, see [[Olga Taussky-Todd]] and [[Richard A. Brualdi]].</ref> The nonnegative eigenvector is often normalized so that the sum of its components is equal to unity; in this case, the eigenvector is the vector of a [[probability distribution]] and is sometimes called a ''stochastic eigenvector''. ''Perron–Frobenius eigenvalue'' and ''dominant eigenvalue'' are alternative names for the Perron root. Spectral projections are also known as ''spectral projectors'' and ''spectral idempotents''. The period is sometimes referred to as the ''index of imprimitivity'' or the ''order of cyclicity''. ==See also== * {{Annotated link|Min-max theorem}} * {{Annotated link|Z-matrix (mathematics)}} * {{Annotated link|M-matrix}} * {{Annotated link|P-matrix}} * {{Annotated link|Routh–Hurwitz matrix}} * [[Metzler matrix]] ([[Quasipositive matrix]]) * {{Annotated link|Positive operator}} * {{Annotated link|Krein–Rutman theorem}} ==Notes== {{reflist|30em|refs= <ref name="Kitchens">{{citation | title=Symbolic dynamics: one-sided, two-sided and countable state markov shifts. | year=1998 | first1=Bruce | last1=Kitchens | url = https://books.google.com/books?id=mCcdC_5crpoC&q=kitchens%20perron%20frobenius&pg=PA16|publisher = Springer | isbn=9783540627388 }}</ref> <ref name="Smith">{{Cite FTP | title = A Spectral Theoretic Proof of Perron–Frobenius| pages = 29–35| volume = 102 | number = 1 | year=2006 | first1=Roger | last1=Smith | url = ftp://emis.maths.adelaide.edu.au/pub/EMIS/journals/MPRIA/2002/pa102i1/pdf/102a102.pdf | server = Mathematical Proceedings of the Royal Irish Academy| url-status = dead| doi=10.3318/PRIA.2002.102.1.29 }}</ref> }} ==References== *{{Citation | last1=Perron | first1=Oskar | title=Zur Theorie der Matrices | doi=10.1007/BF01449896 | year=1907 | journal= [[Mathematische Annalen]] | volume=64 | issue=2 | pages=248–263 | hdl=10338.dmlcz/104432 | s2cid=123460172 | author-link = Oskar Perron | url=http://gdz.sub.uni-goettingen.de/dms/load/toc/?PID=PPN235181684_0064 | hdl-access=free }} *{{citation|first=Georg|last= Frobenius|title=Ueber Matrizen aus nicht negativen Elementen|journal= [[:wikisource:de:Sitzungsberichte der Königlich Preußischen Akademie der Wissenschaften zu Berlin|Sitzungsberichte der Königlich Preussischen Akademie der Wissenschaften]] |date=May 1912|pages=456–477 | author-link = Georg Frobenius }} *{{citation|first=Georg|last= Frobenius|title=Über Matrizen aus positiven Elementen, 1 |journal= Sitzungsberichte der Königlich Preussischen Akademie der Wissenschaften |year=1908|pages=471–476 | author-link = Georg Frobenius}} *{{citation|first=Georg|last= Frobenius|title=Über Matrizen aus positiven Elementen, 2 |journal= Sitzungsberichte der Königlich Preussischen Akademie der Wissenschaften |year=1909|pages=514–518 | author-link = Georg Frobenius }} * {{citation | title= The Theory of Matrices, Volume 2 | orig-year = 1959| year= 2000 | first1=Felix | last1=Gantmacher | publisher = AMS Chelsea Publishing| author1-link=Felix Gantmacher| isbn= 978-0-8218-2664-5| url = https://books.google.com/books?id=cyX32q8ZP5cC&q=preceding%20section&pg=PA53}} (1959 edition had different title: "Applications of the theory of matrices". Also the numeration of chapters is different in the two editions.) *{{citation|title=Google page rank and beyond |year=2006 |first1=Amy |last1=Langville |first2=Carl |last2=Meyer |publisher=Princeton University Press |isbn=978-0-691-12202-1 |doi=10.1007/s10791-008-9063-y |s2cid=7646929 }} *{{citation | title=The Perron–Frobenius theorem and the ranking of football teams | year=1993 | first1=James | last1=Keener | jstor = 2132526| journal = SIAM Review| volume = 35| number =1| pages = 80–93 | doi=10.1137/1035004 }} *{{citation |title = Matrix analysis and applied linear algebra |year = 2000 |first1 = Carl |last1 = Meyer |url = http://www.matrixanalysis.com/Chapter8.pdf |publisher = SIAM |isbn = 978-0-89871-454-8 |url-status = dead |archive-url = https://web.archive.org/web/20100307021652/http://www.matrixanalysis.com/Chapter8.pdf |archive-date = 2010-03-07 }} *{{citation|first=Henryk |last= Minc |title=Nonnegative matrices |isbn=0-471-83966-3 |year=1988 |publisher=John Wiley&Sons,New York}} *{{citation|first=V.|last= Romanovsky|title=Sur les zéros des matrices stocastiques |journal= Bulletin de la Société Mathématique de France |year=1933|pages=213–219 | volume = 61 |doi= 10.24033/bsmf.1206|doi-access=free}} *{{citation|first=Lothar |last= Collatz| author-link = Lothar Collatz|title=Einschließungssatz für die charakteristischen Zahlen von Matrizen |journal= Mathematische Zeitschrift |year=1942|pages=221–226 |doi = 10.1007/BF01180013| volume = 48 | number =1 |s2cid= 120958677}} *{{citation|first=Helmut|last= Wielandt|title=Unzerlegbare, nicht negative Matrizen |journal= Mathematische Zeitschrift |year=1950|pages=642–648 |doi = 10.1007/BF02230720| volume = 52 | number =1 |hdl= 10338.dmlcz/100322|s2cid= 122189604|hdl-access=free}} ===Further reading=== * Abraham Berman, [[Robert J. Plemmons]], ''Nonnegative Matrices in the Mathematical Sciences'', 1994, SIAM. {{isbn|0-89871-321-8}}. * [[Chris Godsil]] and [[Gordon Royle]], ''Algebraic Graph Theory'', Springer, 2001. * A. Graham, ''Nonnegative Matrices and Applicable Topics in Linear Algebra'', John Wiley&Sons, New York, 1987. * R. A. Horn and C.R. Johnson, ''Matrix Analysis'', Cambridge University Press, 1990 * Bas Lemmens and Roger Nussbaum, ''Nonlinear Perron-Frobenius Theory'', Cambridge Tracts in Mathematics 189, Cambridge Univ. Press, 2012. * S. P. Meyn and R. L. Tweedie, [https://web.archive.org/web/20100619010320/https://netfiles.uiuc.edu/meyn/www/spm_files/book.html ''Markov Chains and Stochastic Stability''] London: Springer-Verlag, 1993. {{isbn|0-387-19832-6}} (2nd edition, Cambridge University Press, 2009) * Seneta, E. ''Non-negative matrices and Markov chains''. 2nd rev. ed., 1981, XVI, 288 p., Softcover Springer Series in Statistics. (Originally published by Allen & Unwin Ltd., London, 1973) {{isbn|978-0-387-29765-1}} *{{eom|id=P/p072350|first=D.A. |last=Suprunenko}} (The claim that ''A''<sub>''j''</sub> has order ''n''/''h'' at the end of the statement of the theorem is incorrect.) *{{citation|first=Richard S.|last=Varga|author-link=Richard S. Varga|title=Matrix Iterative Analysis|edition=2nd|publisher=Springer-Verlag|year=2002}}. {{DEFAULTSORT:Perron-Frobenius Theorem}} [[Category:Matrix theory]] [[Category:Theorems in linear algebra]] [[Category:Markov processes]]
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)
Pages transcluded onto the current version of this page
(
help
)
:
Template:Annotated link
(
edit
)
Template:Citation
(
edit
)
Template:Citation needed
(
edit
)
Template:Cite book
(
edit
)
Template:Cite journal
(
edit
)
Template:Clarify
(
edit
)
Template:Eom
(
edit
)
Template:Harvnb
(
edit
)
Template:Harvs
(
edit
)
Template:Isbn
(
edit
)
Template:Main
(
edit
)
Template:Reflist
(
edit
)
Template:Short description
(
edit
)