Let $M$ be a row substochastic matrix, with at least one row having sum less than 1. Also, suppose $M$ is irreducible in the sense of a Markov chain. Is there an easy way to show the largest eigenvalue must be strictly less than 1? I hope that this result is true as stated. I know that Cauchy interlacing theorem gives me $\leq$,
-
2I think you are missing a condition, else you could take $M$ to be the 2 by 2 diagonal matrix with entries 1/2 and 1. The largest eigenvalue of this $M$ is 1. – May 04 '11 at 00:59
-
Sorry I meant to add irreducibility. Thanks for the correction – SKS May 04 '11 at 01:19
5 Answers
A bit late to the game, but I thought of this proof.
Suppose $A$ is an irreducible sub-stochastic matrix and $\lambda$ is the the Perron-Frobenius eigenvalue of $A$ (i.e. $\rho\left(A\right) = \lambda$) with $v$ the corresponding eigenvector normalized such that $\|v\|_{1} = 1$. By the Perron-Frobenius theorem for irreducible non-negative matrices, the entries of $v$ must be positive. Using this, we have the following.
\begin{align} |\lambda| &= \|\lambda v\|_{1} \\ &= \|vA\|_{1} \\ &= \sum_j\sum_k v_jA_{jk} \end{align} Let $\epsilon_j \doteq \frac{1}{N}\left(1 - \sum_{k=1}^N A_{jk}\right)$. If we add $\epsilon_j$ to each element of the $j$th row of A, the row sum will become one. Let $\boldsymbol\epsilon$ be the row vector containing the values of $\{\epsilon_j\}$. \begin{align} |\lambda| &= \sum_j \sum_k v_j\left(A_{jk} + \epsilon_j - \epsilon_j\right) \\ &= \sum_j \sum_k v_j\left(A_{jk} + \epsilon_j\right) -\sum_j \sum_k v_j \epsilon_j \\ &= \left\|v\left(A + \boldsymbol\epsilon^T\mathbf{1}\right)\right\|_1 - N \left(\boldsymbol\epsilon\cdot v\right) \end{align}
We define $\hat A \doteq A + \boldsymbol\epsilon^T\mathbf{1}$ and note that it is a proper stochastic matrix. Since $v$ is positive and $\boldsymbol\epsilon$ is non-negative with at least one positive entry we have $\boldsymbol\epsilon\cdot v > 0$. \begin{align} |\lambda| &= \left\| v \hat A \right\|_{1} - N \left(\boldsymbol\epsilon\cdot v\right) \\ &= 1 - N \left(\boldsymbol\epsilon\cdot v\right) \\ &< 1 \end{align}
- 191
- 1
- 4
-
I might have missed something but I think your argument requires the matrix to be primitive, which may not be the case here. – Aiden Aug 26 '15 at 05:29
-
@Aiden Do you mean that being primitive is a necessary condition to invoke the Perron-Frobenius Theorem? That was not my understanding, but I could be wrong. – Daniel Johnson Aug 30 '15 at 15:35
-
I'm sorry your proof is right. But it is just the assertion that "the entries of $v$ must be positive" is a bit too strong. I think it should be we can choose it to be positive to be more rigorous. The "must" leads me to think about period 1 thing. Anyway I'am just being captious and making excuses honestly. your proof is a beautiful one. – Aiden Aug 30 '15 at 16:17
-
@Aiden The Perron-Frobenius theorem for irreducible non-negative matrices states, among other things, that the eigenvector in question can be chosen positive and $\ell^1$-normalised (in which case it's called the "Frobenius vector") . Of course, you can make it negative by multiplying by $-1$. The point is that it belongs to a 1d-eigenspace intersecting all coordinate hyperplanes (i.e., all components are non-zero). I don't see how this might be related to possible aperiodicity or primitivity of the given matrix. – paperskilltrees Oct 15 '21 at 05:11
You can try completing your matrix to a Markov chain, adding a self loop at the additional state. The new Markov chain is irreducible and aperiodic and so has a unique stationary distribution, which is concentrated on the additional state. It is also the unique eigenvector with eigenvalue at least $1$.
Now take a purported eigenvector for your original matrix with eigenvalue $1$, and add a zero coordinate. The result is an eigenvector for the Markov chain, contradicting the properties we listed above.
In effect, you have a Markov chain with an invisible absorbing state, which is actually reachable from any other state. This ensures that in the long run the state will be reached, and so applying your matrix lots of times on any given vector will yield the zero vector. So all eigenvalues must be less than 1 in magnitude.
- 57,953
-
I'm not quite seeing that the sub stochastic eigenvector, when extended with a zero coordinate becomes an eigenvector of the new markov chain, particularly when looking at what happens to the last coordinate of the eigenvector upon multiplication – SKS May 04 '11 at 22:25
-
The last coordinate is zero in both "input" and "output", since the corresponding row is $0,\ldots,0,1$ (the last entry corresponding to the added state). – Yuval Filmus May 04 '11 at 23:48
-
1Something is bothering me about your argument. For instance, if you add an absorbing state the resulting chain is certainly not irreducible. – May 05 '11 at 00:04
-
@Byron I guess you're right about that! But it's still true that there is only one stationary probability since wherever you start, you'll almost surely reach the absorbing vertex. This is basically the argument in your answer. – Yuval Filmus May 05 '11 at 01:25
This is essentially Yuval's probabilistic argument with the probability removed. The goal is to show that powers of $M$ converge to zero.
For any state $i$ and integer $n\geq 0$, let $r^n_i=\sum_k M^n_{i k}$ denote the $i$th row sum of $M^n$. For $n=1$, we write $r_i$ rather than $r^1_i$. Since $M$ is substochastic we have $0\leq r^n_i\leq 1$.
Let $k^*$ be an index with $r_{k^*}<1$, and note that for $n\geq 1$
$$r^n_{k^*}=\sum_k M_{k^* k}\ r_k^{n-1}\leq \sum_k M_{k^* k} =r_{k^*}<1.$$
By irreducibility, for any $i$, there is an $m$ with $M_{i k^*}^m>0$.
In fact, if $M$ is $N\times N$ matrix, and $i\neq k^*$ then we can
assume $m<N$. (Take the shortest path from $i$ to $k^*$ with positive "probability").
Since $M_{i k}^m$ puts positive weight on the index $k=k^*$, we have
$$r^N_i=\sum_k M^m_{i k}\ r^{N-m}_k < r^m_i \leq 1.$$
That is, every row sum of $M^N$ is strictly less than one. Now you can show that $M^{jN}\to 0$ as $j\to \infty$ and this shows that $M^N$ (and hence $M$) cannot have any eigenvalue with modulus 1.
Sorry for the almost decade-old necropost; I saw this question and my interest was piqued since this is something tangential to a topic I worked on. I apologize in advance for the self-plug.
Let $M$ be a substochastic matrix. Let's call $M$ irreducibly substochastic if it is irreducible and has at least one row that sums to strictly less than one (this is exactly the condition stated in your post). The other answers to this post establish \begin{equation} M \text{ is irreducibly substochastic} \implies \rho(M) < 1. \end{equation} We can actually establish a stronger "iff" by relaxing the requirement on the left hand side: \begin{equation} \label{IFF} M \text{ is weakly chained substochastic} \iff \rho(M) < 1 \tag{IFF}. \end{equation}
The definition of weakly chained substochastic is given below.
Definition. $M$ is weakly chained substochastic if and only if for each row $i$ at least one of the following is true:
- row $i$ sums to less than one (i.e., $\sum_j M_{ij} < 1$) or
- there is a walk $i \rightarrow i_1 \rightarrow \cdots \rightarrow i_k$ in the adjacency graph of $M$ s.t. row $i_k$ sums to less than one.
Any irreducibly substochastic matrix is trivially a weakly chained substochastic matrix since it has at least one row that sums to less than one (call it $j$) and for any row $i$ we can, by irreducibility, construct a walk from $i$ to $j$.
You can find a proof of \eqref{IFF} in Corollary 2.6 in Azimzadeh, Parsiad. "A fast and stable test to check if a weakly diagonally dominant matrix is a nonsingular M-matrix." Mathematics of Computation 88.316 (2019): 783-800. [arXiv].
Example. The matrix $$ M = \begin{pmatrix}0.1 & 0.9 & 0 \\ 0 & 0.2 & 0.8 \\ 0 & 0 & 0.3 \end{pmatrix} $$ is substochastic. It is, however, reducible. Note that $1 \rightarrow 2 \rightarrow 3$ and $2 \rightarrow 3$ are walks in the adjacency graph of $M$. Moreover, row $3$ sums to less than one. As such, we can apply the above theorem to conclude that $\rho(M) < 1$.
Example. On the other hand, the adjacency graph of the matrix $$ M = \begin{pmatrix}0.1 & 0.9 & 0 \\ 0.8 & 0.2 & 0 \\ 0 & 0 & 0.3 \end{pmatrix} $$ has two disjoint strongly connected components $\{1, 2\}$ and $\{3\}$. The first component does not have a row that sums to less than one, and hence $\rho(M) = 1$.
- 25,738
Here is a simple proof. First, for a (sub-)stochastic matrix we have $\rho(\boldsymbol{A})\leq1$, which follows, e.g., from $\rho(\boldsymbol{A})\leq\|\boldsymbol{A}\|_\infty$. Thus, all we want to show is $\rho(\boldsymbol{A})\neq1$. Second, the "substochasticity" can be written as $\boldsymbol{A}\mathsf{1}\leq\mathsf{1}$, $\boldsymbol{A}\mathsf{1}\neq\mathsf{1}$. If $\boldsymbol{A}$ is irreducible, then we apply the following clever proposition with $\alpha=1$ and $\boldsymbol{z}=\mathsf{1}$. This does the job.
Proposition. Let $\boldsymbol{A}\geq0$ be irreducible and $r:=\rho(\boldsymbol{A})$. Then $\boldsymbol{A}\boldsymbol{z}\leq\alpha\boldsymbol{z}$, $\boldsymbol{A}\boldsymbol{z}\neq\alpha\boldsymbol{z}$ together imply $\alpha\neq r$.
Proof. Let $\boldsymbol{c}:=\left(\boldsymbol{A}-\alpha\boldsymbol{I}\right)\boldsymbol{z}$ then $\boldsymbol{A}\boldsymbol{z}\leq\alpha\boldsymbol{z}$ and $\boldsymbol{A}\boldsymbol{z}\neq\alpha\boldsymbol{z}$ become $\boldsymbol{c}\leq0$ and $\boldsymbol{c}\neq0$. Since $\boldsymbol{A}\geq0$ is irreducible, there exists the left Frobenius vector $\boldsymbol{\phi}>0$, such that $\boldsymbol{\phi}^{T}(\boldsymbol{A}-r\boldsymbol{I})=\boldsymbol{0}$ (by the Perron-Frobenius theorem). Note that $\boldsymbol{\phi}>0$ and $\boldsymbol{c}\leq0,\boldsymbol{c}\neq0$ imply $\boldsymbol{\phi}^{T}\boldsymbol{c}\neq0$. If $\alpha=r$ then $\boldsymbol{\phi}^{T}(\boldsymbol{A}-r\boldsymbol{I})\boldsymbol{z}=\boldsymbol{\phi}^{T}\boldsymbol{c}=0$, which contradicts $\boldsymbol{\phi}^{T}\boldsymbol{c}\neq0$. Therefore, $\alpha\neq r$. $\blacksquare$
Background. This proposition is similar to Example 8.3.1 on p.674 of Carl Meyer's "Matrix Analysis and Applied Linear Algebra", where Meyer makes and proves a different but somewhat similar statement, which he later applies to substochastic matrices. I couldn't figure out how Meyer's statement on p.674 was relevant, and I believe he just made a mistake, which I think is corrected by replacing his statement with my proposition.
- 755