Convergence of indirect methods

7.3. Convergence of indirect methods#

We have seen that both the Jacobi and Gauss-Seidel methods are iterated until the estimates of the solution converge to a given tolerance. In Example 7.1 required 49 iterations to converge to the solution of a system of linear equations whereas in Example 7.2 only required 20 iterations to converge to the solution for the same system.

Not all indirect methods will converge for a given system of linear equations, we can establish whether a method will be convergent using the theorem below. Let \(\mathbf{e}^{(k)} = \|\mathbf{x} - \mathbf{x}^{(k)}\|\) be the error between the exact solution \(\mathbf{x}\) and the estimate \(\mathbf{x}^{(k)}\). The error from one estimate to the next is updated using the iteration matrix for the method

\[ \mathbf{e}^{(k+1)} = T\mathbf{e}^{(k)}. \]

We can write the first error, \(\mathbf{e}^{(0)}\) as a linear combination of some vectors \(\mathbf{v}_i\)

\[ \mathbf{e}^{(0)} =\alpha_1 \mathbf{v}_1 +\alpha_2 \mathbf{v}_2 +\cdots +\alpha_n \mathbf{v}_n =\sum_{i=1}^n \alpha_i \mathbf{v}_i, \]

where \(\alpha_i\) are scalars. If \(\mathbf{v}_i\) are the eigenvalues of the iteration matrix \(T\) with eigenvalues \(\lambda_i\) so \(T\mathbf{v}_i = \lambda_i \mathbf{v}_i\) then iterating \(\mathbf{e}^{(k)}\) gives

\[\begin{split} \begin{align*} \mathbf{e}^{(1)} &= T\mathbf{e}^{(0)} =T\left(\sum_{i=1}^n \alpha_i \mathbf{v}_i \right)=\sum_{i=1}^n \alpha_i T\mathbf{v}_i = \sum_{i=1}^n \alpha_i \lambda_i \mathbf{v}_i , \\ \mathbf{e}^{(2)} &=T\mathbf{e}^{(1)} =T\left(\sum_{i=1}^n \alpha_i \lambda_i \mathbf{v}_i \right)=\sum_{i=1}^n \alpha_i \lambda_i T\mathbf{v}_i =\sum_{i=1}^n \alpha_i \lambda_i^2 \mathbf{v}_i , \\ &\vdots \\ \mathbf{e}^{(k+1)} &=\sum_{i=1}^n \alpha_i \lambda_i^{k+1} \mathbf{v}_i . \end{align*} \end{split}\]

If \(|\lambda_1|>\lambda_2, \lambda_3, \ldots \lambda_n\) then

\[ \mathbf{e}^{(k+1)} =\alpha_1 \lambda_1^{k+1} \mathbf{v}_1 +\sum_{i=2}^n \alpha_i \lambda_i^{k+1} \mathbf{v}_i =\lambda_1^{k+1} \left(\alpha_1 \mathbf{v}_1 +\sum_{i=2}^n \alpha_i \mathbf{v}_i {\left(\frac{\lambda_i }{\lambda_1 }\right)}^{k+1} \right) \]

and since \(\dfrac{\lambda_i}{\lambda_ 1} < 1\) then

\[ \begin{align*} \lim_{k\to \infty } \mathbf{e}^{(k+1)} =\alpha_1 \lambda_1^{k+1} \mathbf{v}_1 . \end{align*} \]

This means that as the number of iterations increases, the error varies by a factor of \(\lambda_1^{k+1}\) where \(\lambda_1\) is the largest eigenvalue of \(T\) which is known as the spectral radius.

Definition 7.4 (Spectral radius)

The spectral radius of \(A\) denoted by \(\rho(A)\) is

(7.8)#\[ \rho(A) = \max_i(| \lambda_i |). \]

where \(\lambda_i\) are the eigenvalues of \(A\).

Theorem 7.1 (Convergence criteria for an indirect method)

The spectral radius of \(T\) gives us the following information about an indirect method

  • If \(\rho (T) > 1\) then the errors will increase over each iteration, therefore for an indirect method to converge to the solution we require \(\rho (T)< 1\).

  • The smaller the value of \(\rho (T)\) the faster the errors will tend to zero.

Example 7.3

Show that the Jacobi and Gauss-Seidel methods are convergent of the system of linear equations from Example 7.1 (shown below)

\[\begin{split} \begin{align*} 4x_1 +3x_2 &=-2, \\ 3x_1 +4x_2 -x_3 &=-8, \\ -x_2 +4x_3 &=14. \end{align*} \end{split}\]

Solution

The coefficient matrix for this linear system is

\[\begin{split} A = \begin{pmatrix} 4 & 3 & 0 \\ 3 & 4 & -1 \\ 0 & -1 & 4 \end{pmatrix}. \end{split}\]

so

\[\begin{split} \begin{align*} L &= \begin{pmatrix} 0 & 0 & 0 \\ 3 & 0 & 0 \\ 0 & -1 & 0 \end{pmatrix}, & D &= \begin{pmatrix} 4 & 0 & 0 \\ 0 & 4 & 0 \\ 0 & 0 & 4 \end{pmatrix}, & U &= \begin{pmatrix} 0 & 3 & 0 \\ 0 & 0 & -1 \\ 0 & 0 & 0 \end{pmatrix}. \end{align*} \end{split}\]

The iteration matrices for the Jacobi and Gauss-Seidel methods are given in equations (7.4) and (7.7) which for this system are

\[\begin{split} \begin{align*} T_J &= -D^{-1} ( L + U) \\ &= -\begin{pmatrix} 4 & 0 & 0 \\ 0 & 4 & 0 \\ 0 & 0 & 4 \end{pmatrix} ^{-1} \left( \begin{pmatrix} 0 & 0 & 0 \\ 3 & 0 & 0 \\ 0 & -1 & 0 \end{pmatrix} + \begin{pmatrix} 0 & 3 & 0 \\ 0 & 0 & -1 \\ 0 & 0 & 0 \end{pmatrix} \right) \\ &= - \begin{pmatrix} \frac{1}{4} & 0 & 0 \\ 0 & \frac{1}{4} & 0 \\ 0 & 0 & \frac{1}{4} \end{pmatrix} \begin{pmatrix} 0 & 3 & 0 \\ 3 & 0 & -1 \\ 0 & -1 & 0 \end{pmatrix} \\ &= \begin{pmatrix} 0 & -\frac{3}{4} & 0 \\ -\frac{3}{4} & 0 & \frac{1}{4} \\ 0 & \frac{1}{4} & 0 \end{pmatrix}, \\ T_{GS} &= - (L + D)^{-1} U \\ &= -\left( \begin{pmatrix} 0 & 0 & 0 \\ 3 & 0 & 0 \\ 0 & -1 & 0 \end{pmatrix} + \begin{pmatrix} 4 & 0 & 0 \\ 0 & 4 & 0 \\ 0 & 0 & 4 \end{pmatrix} \right)^{-1} \begin{pmatrix} 0 & 3 & 0 \\ 0 & 0 & -1 \\ 0 & 0 & 0 \end{pmatrix} \\ &= - \begin{pmatrix} 3 & 0 & 0 \\ 3 & 4 & 0 \\ 0 & -1 & 4 \end{pmatrix}^{-1} \begin{pmatrix} 0 & 3 & 0 \\ 0 & 0 & -1 \\ 0 & 0 & 0 \end{pmatrix} \\ &= \begin{pmatrix} -\frac{1}{4} & 0 & 0 \\ \frac{3}{16} & \frac{1}{4} & 0 \\ \frac{3}{64} & -\frac{1}{16} & \frac{1}{4} \end{pmatrix} \begin{pmatrix} 0 & 3 & 0 \\ 0 & 0 & -1 \\ 0 & 0 & 0 \end{pmatrix} \\ & = \begin{pmatrix} 0 & -\frac{3}{4} & 0 \\ 0 & \frac{9}{16} & \frac{1}{4} \\ 0 & \frac{9}{64} & \frac{1}{16} \end{pmatrix} \end{align*} \end{split}\]

Calculating the spectral radius for these iteration matrices gives \(\rho(T_J )=0.7906\) and \(\rho (T_{GS})=0.625\) which are both less than 1 so both of these methods are convergent for this system. Furthermore, the Gauss-Seidel method will converge faster than the Jacobi method since it has a smaller spectral radius.