Numerical Solution of Linear Equations & Power Method

Transcription

Numerical Solution of Linear Equations & Power Method
Solution of System of Linear Equations
&
Eigen Values and Eigen Vectors
P. Sam Johnson
March 30, 2015
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
1/43
Overview
Linear systems
Ax = b
(1)
occur widely in applied mathematics.
They occur as direct formulations of “real world” problems ; but more
often, they occur as a part of numerical analysis.
There are two general categories of numerical methods for solving (1).
Direct methods are methods with a finite number of steps ; and they end
with the exact solution provided that all arithmetic operations are exact.
Iterative methods are used in solving all types of linear systems, but they
are most commonly used for large systems.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
2/43
Overview
Iterative methods are faster than the direct methods. Even the round-off
errors in iterative methods are smaller.
In fact, iterative method is a self correcting process and any error made
at any stage of computation gets automatically corrected in the
subsequent steps.
We start with a simple example of a linear system and we discuss
consistency of system of linear equations, using the concept of
determinant.
Two iterative methods – Gauss-Jacobi and Gauss-Seidel methods are
discussed, to solve a given linear system numerically.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
3/43
Overview
Eigen values are a special set of scalars associated with a linear system of
equations. The determination of the eigen values and eigen vectors of a
system is extremely important in physics and engineering, where it is
equivalent to matrix diagonalization and arises in such common
applications as stability analysis, the physics of rotating bodies, and small
oscillations of vibrating systems, to name only a few.
Each eigen value is paired with a corresponding vector, called eigen
vector.
We discuss properties of eigen values and eigen vectors in the lecture. Also
we give a method, called Rayeigh’s Power Method, to find out the
largest eigen value in magnitude (also called, dominant eigen value). The
special advantage of the power method is that the eigen vector corresponds
to the dominant eigen value is also calculated at the same time.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
4/43
An Example
A shopkeeper offers two standard packets because he is convinced that
north indians each more wheat than rice and south indians each more rice
than wheat. Packet one P1 : 5kg wheat and 2kg rice ; Packet two P2 :
2kg wheat and 5kg rice. Notation. (m, n) : m kg wheat and n kg rice.
Suppose I need 19kg of wheat and 16kg of rice. Then I need to buy x
packets of P1 and y packets of P2 so that x(5, 2) + y (2, 5) = (10, 16).
Hence I have to solve the following system of linear equations
5x + 2y
= 10
2x + 5y
= 16.
Suppose I need 34 kg of wheat and 1 kg of rice. Then I must buy 8
packets of P1 and −3 packets of P2 . What does this mean? I buy 8
packets of P1 and from these I make three packets of P2 and give them
back to the shopkeeper.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
5/43
Consistency of System of Linear Equations – Graphically
(1 Equation, 2 Variables)
The solution of the equation
ax + by = c
is the set of all points satisfy the equation forms a straight line in the
plane through the point (c/b, 0) and with slope −a/b.
Two lines
parallel – no solution
intersect – unique solution
same – infinitely many solutions.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
6/43
Consistency of System of Linear Equations
Consider the system of m linear equations in n unkowns
a11 x1 + a12 x2 + · · · + a1n xn = b1
a21 x1 + a22 x2 + · · · + a2n xn = b2
..
.
.
.
. + .. + · · · + ..
= ..
am1 x1 + am2 x2 + · · · + amn xn = bm .
Its matrix form is

a11
 a21

 ..
 .
a12
a22
..
.
...
...
  
x1
b1
a1n
 x2   b2
a2n 
  
..   ..  =  ..
.  .   .
...
am1 am2 . . . amn
xm





bm .
That is, Ax = b, where A is called the coefficient matrix.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
7/43
To determine whether these equations are consistent or not, we find the
ranks of the coefficient matrix A and the augmented matrix


a11 a12 . . . a1n b1
 a21 a22 . . . a2n b2 


K = .
..
..
..  = [A b].
 ..
.
···
.
. 
am1 am2
. . . amn bm
We denote the rank of A by r (A).
1. r (A) 6= r (K ), then the linear system Ax = b is inconsistent and has
no solution.
2. r (A) = r (K ) = n, then the linear system Ax = b is consistent and
has a unique solution.
3. r (A) = r (K ) < n, then the linear system Ax = b is consistent and
has an infinite number of solutions.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
8/43
Solution of Linear Simultaneous Equations
Simultaneous linear equations occur quite often in engineering and science.
The analysis of electronic circuits consisting of invariant elements, analysis
of a network under sinusoidal steady-state conditions, determination of the
output of a chemical plant, finding the cost of chemical reactions are some
of the problems which depend on the solution of simultaneous linear
algebraic equations, the solution of such equations can be obtained by
direct or iterative methods (successive approximation methods).
Some direct methods are as follows:
1. Method of determinants – Cramer’s rule
2. Matrix inversion method
3. Gauss elimination method
4. Gauss-Jordan method
5. Factorization (triangulization) method.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
9/43
Iterative Methods
Direct methods yield the solution after a certain amount of computation.
On the other hand, an iterative method is that in which we start from an
approximation to the true solution and obtain better and better
approximations from a computation cycle repeated as often as may be
necessary for achieving a desired accuracy.
Thus in an iterative method, the amount of computation depends on the
degree of accuracy required.
For large systems, iterative methods are faster than the direct methods.
Even the round-off errors in iterative methods are smaller. In fact,
iteration is a self correcting process and any error made at any stage of
computation gets automatically corrected in the subsequent steps.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
10/43
Diagonally Dominant
Definition
An n × n matrix A is said to be diagonally dominant if, in each row, the
absolute value of each leading diagonal element is greater than or equal to
the sum of the absolute values of the remaining elements in that row.


10 −5 −2
The matrix A =  4 −10 3  is a diagonally dominant and the
1
6
10


2 3 −1
matrix A = 5 8 −4 is not a diagonally dominant matrix.
1 1 1
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
11/43
Diagonal System
Definition
In the sytem of simultaneous linear equations in n unknowns AX = B, if A
is diagonally dominant, then the system is said to be diagonal system.
Thus the system of linear equations
a1 x + b1 y + c1 z
= d1
a2 x + b2 y + c2 z
= d2
a3 x + b3 y + c3 z
= d3
is a diagonal system if
|a1 | ≥ |b1 | + |c1 |
|b2 | ≥ |a2 | + |c2 |
|c3 | ≥ |a3 | + |b3 |.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
12/43
The process of iteration in solving AX = B will converge quickly if the
cofficient matrix A is diagonally dominant.
If the coefficient matrix is not diagonally dominant we must rearrange the
equations in such a way that the resulting coefficient matrix becomes
dominant, if possible, before we apply the iteration method.
Exercises
1. Is the system of equations diagonally dominant? If not, make it diagonally dominant.
3x + 9y − 2z
=
10
4x + 2y + 13z
=
19
4x − 2y + z
=
3.
x + 6y − 2z
=
5
4x + y + z
=
6
−3x + y + 7z
=
5.
2. Check whether the system of equations
is a diagonal system. If not, make it a diagonal system.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
13/43
Jacobi Iteration Method (also known as Gauss-Jacobi)
Simple iterative methods can be designed for systems in which the
coefficients of the leading diagonal are large as comparted to others.
We now describe three such methods. Gauss-Jacobi Iteration Method
Consider the system of linear equations
a1 x + b1 y + c1 z
= d1
a2 x + b2 y + c2 z
= d2
a3 x + b3 y + c3 z
= d3 .
If a1 , b2 , c3 are large as compared to other coefficients (if not, rearrange
the given linear equations), solve for x, y , z respectively. That is, the
cofficient matrix is diagonally dominant.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
14/43
Then the system can be written as
x
=
y
=
z
=
1
(d1 − b1 y − c1 z)
a1
1
(d2 − a2 x − c2 z)
b2
1
(d3 − a3 x − b3 y ).
c3
(2)
Let us start with the initital approximations x0 , y0 , z0 for the values of
x, y , z repectively. In the absence of any better estimates for x0 , y0 , z0 ,
these may each be taken as zero.
Substituting these on the right sides of (2), we gt the first approximations.
This process is repeated till the difference between two consecutive
approximations is negligible.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
15/43
Exercises
3. Solve, by Jacobi’s iterative method, the equations
20x + y − 2z
=
17
3x + 20y − z
=
−18
2x − 3y + 20z
=
25.
4. Solve, by Jacobi’s iterative method correct to 2 decimal places, the equations
10x + y − z
=
11.19
x + 10y + z
=
28.08
−x + y + 10z
=
35.61.
5. Solve the equations, by Gauss-Jacobi iterative method
P. Sam Johnson (NITK)
10x1 − 2x2 − x3 − x4
=
3
−2x1 + 10x2 − x3 − x4
=
15
−x1 − x2 + 10x3 − 2x4
=
27
−x1 − x2 − 2x3 + 10x4
=
−9.
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
16/43
Gauss-Seidel Iteration Method
This is a modification of Jacobi’s Iterative Method.
Consider the system of linear equations
a1 x + b1 y + c1 z
= d1
a2 x + b2 y + c2 z
= d2
a3 x + b3 y + c3 z
= d3 .
If a1 , b2 , c3 are large as compared to other coefficients (if not, rearrange
the given linear equations), solve for x, y , z respectively.
Then the system can be written as
P. Sam Johnson (NITK)
x
=
y
=
z
=
1
(d1 − b1 y − c1 z)
a1
1
(d2 − a2 x − c2 z)
b2
1
(d3 − a3 x − b3 y ).
c3
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
(3)
March 30, 2015
17/43
Here also we start with the initial approximations x0 , y0 , z0 for x, y , z
respectively, (each may be taken as zero).
Subsituting y = y0 , z = z0 in the first of the equations (3), we get
x1 =
1
(d1 − b1 y0 − c1 z0 ).
a1
Then putting x = x1 , z = z0 in the second of the equations (3), we get
y1 =
1
(d2 − a2 x1 − c2 z0 ).
b2
Next subsituting x = x1 , y = y1 in the third of the equations (3), we get
z1 =
1
(d3 − a3 x1 − b3 y1 )
c3
and so on.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
18/43
That is, as soon as a new approximation for an unknown is found, it is
immediately used in the next step.
This process of iteration is repeatedly till the values of x, y , z are obtained
to desired degree of accuracy.
Since the most recent approximations of the unknowns are used which
proceeding to the next step, the convergence in the Gauss-Seidel
method is twice as fast as in Gauss-Jacobi method.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
19/43
Convergence Criteria
The condition for convergence of Gauss-Jacobi and Gauss-Seidel methods
is given by the following rule:
The process of iteration will converge if in each equation of the system,
the absolute value of the largest co-efficient is greater than the sum of the
absolute values of all the remaining coefficients.
That is, if the coefficient matrix is a diagonal dominant matrix, then the
process of iteration will converge.
For a given system of linear equations, before finding approximations by
using Gauss-Jacobi / Gauss-Seidel methods, convergence criteria should be
verified. If the coefficient matrix is a not diagonal dominant matrix,
rearrange the equations, so that the new coefficient matrix is diagonal
dominant. Note that all systems of linear equations are not diagonal
dominant.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
20/43
Exercises
6. Apply Gauss-Seidel iterative method to solve the equations
20x + y − 2z
=
17
3x + 20y − z
=
−18
2x − 3y + 20z
=
25.
7. Solve the equations, by Gauss-Jacobi and Gauss-Seidel methods (and compare the values)
27x + 6y − z
=
85
x + y + 54z
=
110
6x + 15y + 2z
=
72.
8. Apply Gauss-Seidel iterative method to solve the equations
P. Sam Johnson (NITK)
10x1 − 2x2 − x3 − x4
=
3
−2x1 + 10x2 − x3 − x4
=
15
−x1 − x2 + 10x3 − 2x4
=
27
−x1 − x2 − 2x3 + 10x4
=
−9.
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
21/43
Relaxation Method
Consider the system of linear equations
a1 x + b1 y + c1 z
= d1
a2 x + b2 y + c2 z
= d2
a3 x + b3 y + c3 z
= d3 .
We define the residuals Rx , Ry , Rz by the relations
P. Sam Johnson (NITK)
Rx
= d1 − a1 x − b1 y − c1 z
Ry
= d2 − a2 x − b2 y − c2 z
Rz
= d3 − a3 x − b3 y − c3 z.
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
(4)
March 30, 2015
22/43
To start with we assume that x = y = z = 0 and calculate the initial
residuals.
Then the residuals are reduced step by step, by giving increments to
the variables.
We note from the equations (4) that if x is increased by 1 (keeping y and
z constant), Rx , Ry and Rz decrease by a1 , a2 , a3 respectively. This is
shown in the following table along with the effects on the residuals when y
and z are given unit increments.
δx = 1
δy = 1
δz = 1
δRx
−a1
−b1
−c1
δRy
−a2
−b2
−c2
δRz
−a3
−b3
−c3
The table is the negative of transpose of the coefficient matrix.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
23/43
At each step, the numerically largest residual is reduced to almost zero.
How can one reduce a particular residual?
To reduce a particular residual, the value of the corresponding variable is
changed.
For example, to reduce Rx by p, x should be increased by p/a1 .
When all the residuals have been reduced to almost zero, the increments
in x, y , z are added separately to give the desired solution.
Verification Process : The residuals are not all negligible when the
computed values of x, y , z are substituted in (4), then there is some
mistake and the entire process should be rechecked.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
24/43
Convergence of the Relaxation Method
Relaxation method can be applied successfully only if there is at least one
row in which diagonal element of the coefficient matrix dominate the
other coefficients in the corresponding row.
That is,
|a1 | ≥ |b1 | + |c1 |
|b2 | ≥ |a2 | + |c2 |
|c3 | ≥ |a3 | + |b3 |
should be valid for at least one row.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
25/43
Exercises
9. Solve by relaxation method, the equations
9x − 2y + z
= 50
x + 5y − 3z
= 18
−2x + 2y + 7z
= 19.
10. Solve by relaxation method, the equations
P. Sam Johnson (NITK)
10x − 2y − 3z
= 205
−2x + 10y − 2z
= 154
−2x − y + 10z
= 120.
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
26/43
Eigen Values and Eigen Vectors
For any n × n matrix A, consider the polynomial
λ − a11 −a12
−a21 λ − a22
χA (λ) := |λI − A| = ···
···
−an1
−an2
···
···
···
···
−a1n
−a2n
···
λ − ann
.
(5)
Clearly this is a monic polynomial of degree n. By the fundamental
theorem of algebra, χ(A) has exactly n (not necessarily distinct) roots.
χA (λ)
χA (λ) = 0
the roots of χA (λ)
distinct roots of χA (λ)
P. Sam Johnson (NITK)
the
A
the
the
the
characteristic polynomial of
characteristic equation of A
characteristic roots of A
spectrum of A
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
27/43
Few Observations on Characteristic Polynomials
The constant terms and the coefficient of λn−1 in χA (λ) are (−1)n |A|
and tr (A).
The sum of the characteristic roots of A is tr (A) and the product of
the characteristic roots of A is |A|.
Since λI − AT = (λI − A)T , characteristic polynomials of A and AT
are the same.
Since λI − P −1 AP = P −1 (λI − A)P, similar matrices have the same
characteristic polynomials.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
28/43
Eigen Values and Eigen Vectors
Definition
A complex number λ is an eigen value of A if there exists x 6= 0 in Cn
such that Ax = λx. Any such (non-zero) x is an eigen vector of A
corresponding to the eigen value λ.
When we say that x is an eigen vector of A we mean that x is an eigen
vector of A corresponding to some eigen value of A.
λ is an eigen value of A iff the system (λI − A)x = 0 has a non-trivial
solution.
λ is a characteristic root of A iff λI − A is singular.
Theorem
A number λ is an eigen value of A iff λ is a characteristic root of A.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
29/43
The preceding theorem shows that eigen values are the same as
characteristic roots. However, by ‘the characteristic roots of A’ we mean
the n roots of the characteristic polynomial of A whereas ‘the eigen values
of A’ would mean the distinct characteristic roots of A.
Equivalent names:
Eigen values
Eigen vectors
P. Sam Johnson (NITK)
proper values, latent roots, etc.
characteristic vectors, latent vectors, etc.
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
30/43
Properties of Eigen Values
1. The sum and product of all eigen values of a matrix A is the sum of
principal diagonals and determinant of A, respectively. Hence a
matrix is not invertible if and only if it has a zero eigen value.
2. The eigen values of A and AT (the transpose of A) are same.
3. If λ1 , λ2 , . . . , λn are eigen values of A, then
(a)
(b)
(c)
(d)
1/λ1 , 1/λ2 , . . . , 1/λn are eigen values of A−1 .
kλ1 , kλ2 , . . . , kλn are eigen values of kA.
kλ1 + m, kλ2 + m, . . . , kλn + m are eigen values of kA + mI.
λp1 , λp2 , . . . , λpn are eigen values of Ap , for any positive integer p.
4. The transformation of A by a non-singular matrix P to P −1 AP is
called a similarity transformation.
Any similarity transformation applied to a matrix leaves its eigen
values unchanged.
5. The eigen values of a real symmetric matrix are real.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
31/43
Cayley - Hamilton Theorem
Theorem
For every matrix A, the characteristic polynomial of A annihilates A.
That is, every matrix satisfies its own characteristic equation.
Main uses of Cayley-Hamilton theorem
1. To evaluate large powers A.
2. To evaluate a polynomial in A with large degree even if A is singular.
3. To express A−1 as a polynomial in A whereas A is non-singular.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
32/43
Dominant Eigen Value
We have seen that the eigen values of an n × n matrix A are obtained by
solving its characteristic equation
λn + an−1 λn−1 + cn−2 λn−2 + · · · + c0 = 0.
For large values of n, polynomial equations like this one are difficult and
time-consuming to solve.
Moreover, numerical techniques for approximating roots of polynomial
equations of high degree are sensitive to rounding errors. We now look at
an alternative method for approximating eigen values.
As presented here, the method can be used only to find the eigen
value of A, that is, largest in absolute value – we call this eigen value the
dominant eigen value of A.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
33/43
Dominant Eigen Vector
Dominant eigen values are of primary interest in many physical
applications.
Let λ1 , λ2 , . . . , and λn be the eigen values of an n × n matrix A.
λ1 is called the dominant eigen value of A if
|λi | > |λj |,
For example, the matrix
i = 2, 3, . . . , n.
1 0
has no dominant eigen value.
0 −1
The eigen vectors corresponding to λ1 are called dominant eigen vectors
of A.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
34/43
Power Method
In many engineering problems, it is required to compute the numerically
largest eigen values (dominant eigen values) and the corresponding eigen
vectors (dominant eigen vectors).
In such cases, the following iterative method is quite convenient which is
also well-suited for machine computations.
Let x1 , x2 , . . . , xn be the eigen vectors corresponding to the eigen values
λ1 , λ2 , . . . , λn .
We assume that the eigen values of A are real and distint with
|λ1 | > |λ2 | > · · · > |λn |.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
35/43
As the vectors x1 , x2 , . . . , xn are linearly independent, any arbitrary column
vector can be written as
x = k1 x1 + k2 x2 + · · · + kn xn .
We now operate A repeatedly on the vector x.
Then
Ax = k1 λ1 x1 + k2 λ2 x2 + · · · + kn λn xn .
Hence
Ax = λ1 k1 x1 + k2
λn
λ2
x2 + · · · + kn
xn .
λ1
λ1
and through iteration we obtain,
r
r λ2
λn
r
A x = λ1 k1 x1 + k2
x2 + · · · + kn
xn
λ1
λ1
(6)
provided λ1 6= 0.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
36/43
Since λ1 is dominant, all terms inside the brackets have limit zero except
the first term.
That is, for large values of r , the vector
r
r
λ2
λn
k1 x1 + k2
x2 + · · · + kn
xn
λ1
λ1
will converge to k1 x1 , that is the eigen vector of λ1 .
The eigenvalue is obtained as
(Ar +1 x)p
n→∞ (Ar x)p
λ1 = lim
p = 1, 2, . . . , n
where the index p signifies the pth component in the corresponding vector.
That is, if we take the ratio of any corresponding components of Ar +1 x
and Ar x, this ratio should therefore have a limit λ1 .
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
37/43
Linear Convergence
Neglecting the terms of λ2 , λ3 , . . . , λn in (6) we get
Ar +1 x
λ2 Ar x
− x1
− x1 =
λ1 k1 λr1
k1 λ1r +1
which shows that the convergence is linear with convergence ratio
λ2 .
λ1 Thus, the convergence is faster than if this ratio is small.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
38/43
Working Rule To Find Dominant Eigen Value
Choose x 6= 0, an arbitrary real vector.
Generally, we choose x as
 
 
1
1
1 , or, 0 .
1
0
In other words, for the purpose of computation, we generally take x with 1
as the first component.
We start with a column vector x which is as near the solution as possible
and evaluate Ax which is written as
λ(1) x (1)
after normalization. This gives the first approximation λ(1) to the eigen
value and x (1) is the eigen vector.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
39/43
Rayeigh’s Power Method
Similarly, we evaluate Ax (2) = λ(2) x (2) which gives the second
approximation. We repeat this process till [x (r ) − x (r −1) ] becomes
negligible.
Then λ(r ) will be the dominant (largest) eigen value and x (r ) , the
corresponding eigen vector (dominant eigen vector).
The iteration method of finding the dominant eigen value of a square
matrix is called the Rayeigh’s power method.
At each stage of iteration, we take out the largest component from y (r ) to
find x (r ) where
y (r +1) = Ax (r ) ,
r = 0, 1, 2, . . . .
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
40/43
To find the Smallest Eigen Value
To find the smallest eigen value of a given matrix A, apply power method
to the matrix A−1 .
Find the dominant eigen value of A−1 and then take is reciprocal.
Exercise
11. Say true or false with justification: If λ is the dominant eigen value of
A and β is the dominant eigen value of A − λI , then the smallest
eigen value of A is λ + β.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
41/43
Exercises
12. Determine the largest eigen value and the corresponding eigen vector
of the matrix
5 4
.
1 2
13. Find the largest eigen value and the corresponding eigen vector of the
matrix


2 −1 0
−1 2 −1
0 −1 2
using power method. Take [1, 0, 0]T as an initial eigen vector.
14. Obtain by power method, the numerically dominant eigen value and
eigen vector of the matrix


15 −4 −3
−10 12 −6 
−20 4 −2.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
42/43
References
1. Richard L. Burden and J. Douglas Faires, Numerical Analysis Theory and Applications, Cengage Learning, Singapore.
2. Kendall E. Atkinson, An Introduction to Numerical Analysis, Wiley
India.
3. David Kincaid and Ward Cheney, Numerical Analysis Mathematics of Scientific Computing, American Mathematical
Society, Providence, Rhode Island.
4. S.S. Sastry, Introductory Methods of Numerical Analysis, Fourth
Edition, Prentice-Hall, India.
P. Sam Johnson (NITK)
Solution of System of Linear Equations & Eigen Values and Eigen Vectors
March 30, 2015
43/43