Iterative methods
From CFD-Wiki
m (Reverted edits by AcelcAceld (Talk) to last version by Jasond) |
|||
(4 intermediate revisions not shown) | |||
Line 1: | Line 1: | ||
- | + | We seek the solution to the linear system of equations <br> | |
- | :<math> | + | :<math> Ax = b </math> <br> |
- | After ''k'' iterations we obtain an | + | Iterative methods, unlike direct methods, generate a sequence of approximate solutions to the system that (hopefully) converges to the exact solution. After ''k'' iterations, we obtain an approximation to the exact solution as: <br> |
- | :<math> Ax^{(k)} = | + | :<math> Ax^{(k)} = b - r^{(k)}, </math><br> |
where <math> r^{(k)} </math> is the residual after ''k'' iterations. <br> | where <math> r^{(k)} </math> is the residual after ''k'' iterations. <br> | ||
- | Defining | + | Defining <br> |
:<math> \varepsilon ^{(k)} = x - x^{(k)} </math> <br> | :<math> \varepsilon ^{(k)} = x - x^{(k)} </math> <br> | ||
- | as the difference between the exact and approaximate solution | + | as the difference between the exact and approaximate solution, we obtain <br> |
- | we obtain | + | :<math> A\varepsilon ^{(k)} = r^{(k)}. </math> <br> |
- | :<math> A\varepsilon ^{(k)} = r^{(k)} </math> <br> | + | The purpose of iterations is to drive this residual to zero. |
- | + | ||
===Stationary Iterative Methods=== | ===Stationary Iterative Methods=== | ||
- | Iterative methods that can be expressed in the simple form | + | Iterative methods that can be expressed in the simple form |
+ | |||
:<math> | :<math> | ||
x^{(k+1)} = Bx^{(k)} + c | x^{(k+1)} = Bx^{(k)} + c | ||
- | </math> | + | </math> |
+ | |||
+ | when neither '''B''' nor '''c''' depend upon the iteration count (k), the iterative method is called stationary iterative method. Some of the stationary iterative methods are | ||
- | |||
#Jacobi method | #Jacobi method | ||
#Gauss-Seidel method | #Gauss-Seidel method | ||
#Successive Overrelaxation (SOR) method and | #Successive Overrelaxation (SOR) method and | ||
#Symmetric Successive Overrelaxation (SSOR) method | #Symmetric Successive Overrelaxation (SSOR) method | ||
+ | |||
+ | The convergence of such iterative methods can be investigated using the [[Fixed point theorem]]. | ||
===Nonstationary Iterative Methods=== | ===Nonstationary Iterative Methods=== | ||
Line 35: | Line 38: | ||
#BiConjugate Gradient Stabilized (Bi-CGSTAB) | #BiConjugate Gradient Stabilized (Bi-CGSTAB) | ||
#Chebyshev Iteration | #Chebyshev Iteration | ||
- | |||
- | |||
- | |||
- |
Latest revision as of 14:17, 19 December 2008
We seek the solution to the linear system of equations
Iterative methods, unlike direct methods, generate a sequence of approximate solutions to the system that (hopefully) converges to the exact solution. After k iterations, we obtain an approximation to the exact solution as:
where is the residual after k iterations.
Defining
as the difference between the exact and approaximate solution, we obtain
The purpose of iterations is to drive this residual to zero.
Stationary Iterative Methods
Iterative methods that can be expressed in the simple form
when neither B nor c depend upon the iteration count (k), the iterative method is called stationary iterative method. Some of the stationary iterative methods are
- Jacobi method
- Gauss-Seidel method
- Successive Overrelaxation (SOR) method and
- Symmetric Successive Overrelaxation (SSOR) method
The convergence of such iterative methods can be investigated using the Fixed point theorem.
Nonstationary Iterative Methods
When during the iterations B and c changes during the iterations, the method is called Nonstationary Iterative Method. Typically, constants B and c are computed by taking inner products of residuals or other vectors arising from the iterative method.
Some examples are:
- Conjugate Gradient Method (CG)
- MINRES and SYMMLQ
- Generalized Minimal Residual (GMRES)
- BiConjugate Gradient (BiCG)
- Quasi-Minimal Residual (QMR)
- Conjugate Gradient Squared Method (CGS)
- BiConjugate Gradient Stabilized (Bi-CGSTAB)
- Chebyshev Iteration