📓 CME 302

        • Drawing 2023-10-04 2.excalidraw
        • Drawing 2023-10-04 12.13.49.excalidraw
        • Drawing 2023-10-04 12.26.03.excalidraw
        • Drawing 2023-10-15 18.35.05.excalidraw
        • Drawing 2023-10-15 18.36.19.excalidraw
        • Drawing 2023-10-18 11.37.23.excalidraw
        • GMRES least-squares problem 2023-11-29 10.49.27.excalidraw
        • Gram-Schmidt 2023-10-15 17.54.00.excalidraw
        • Least-squares solution using SVD 2023-10-15 20.28.15.excalidraw
        • MINRES 2023-12-04 08.40.09.excalidraw
        • QR iteration with shift 2023-10-30 11.49.00.excalidraw
        • QR using Givens transformations 2023-10-18 11.41.30.excalidraw
        • QR using Householder transformations 2023-10-18 11.37.36.excalidraw
      • 2024 complete and painless Conjugate Gradient
      • Accelerating convergence using a shift
      • Algorithm for QR iteration with shift
      • Algorithm for the Arnoldi process
      • Algorithm for the Lanczos process
      • All the orthogonality relations in CG
      • Angle between subspaces
      • Applying a Householder transformation
      • Arnoldi process
      • Backward error analysis for LU
      • Bootcamp
      • Brief introduction to Conjugate Gradients
      • Cauchy-Schwarz
      • CG search directions
      • Chebyshev iteration
      • Cholesky factorization
      • Classical iterative methods to solve sparse linear systems
      • Computational cost of Arnoldi and Lanczos
      • Computationally efficient search directions
      • Computing eigenvalues
      • Computing eigenvectors using the Schur decomposition
      • Computing multiple eigenvalues
      • Conditioning of a linear system
      • Conjugate Gradients algorithm
      • Conjugate Gradients code
      • Conjugate Gradients Version 1
      • Connection between Arnoldi and polynomials of A
      • Connection between Krylov subspace and Arnoldi
      • Convergence of classical iterative methods
      • Convergence of GMRES
      • Convergence of Lanczos eigenvalues for symmetric matrices
      • Convergence of Lanczos inner eigenvalues
      • Convergence of the Conjugate Gradients
      • Convergence of the orthogonal iteration
      • Convergence with conjugate steps
      • Deflation in the QR iteration
      • Determinant
      • Diagonalizable matrices
      • Dot product
      • Eigenvalues
      • Eigenvalues cannot be computed exactly
      • Exact shift
      • Existence of LU
      • Existence of the Cholesky factorization
      • Flexible preconditioned conjugate gradient method
      • Floating point arithmetic and unit roundoff error
      • Floating point arithmetic is different from regular arithmetic
      • Floating point numbers
      • Forward and backward error
      • Gauss-Seidel iteration
      • Ghost eigenvalues in the Lanczos process
      • GMRES
      • GMRES algorithm
      • GMRES least-squares problem
      • Gram-Schmidt
      • Hermitian and symmetric matrices
      • Householder transformation
      • Invertible matrix
      • Iterative methods for eigenvalue computation
      • Jacobi iteration
      • Key idea of iterative methods for eigenvalue computation
      • Krylov iterative methods to solve sparse linear systems
      • Krylov methods for sparse systems
      • Krylov subspace
      • Lanczos process
      • Least-squares problems
      • Least-squares solution using QR
      • Least-squares solution using SVD
      • LU algorithm
      • LU and determinant
      • Matrix block operations
      • Matrix-vector and matrix-matrix product
      • Method of normal equation
      • Method of power iteration
      • MINRES
      • Motivation of iterative methods for eigenvalue computation
      • Operator and matrix norms
      • Optimal step size
      • Orthogonal iteration
      • Orthogonal iteration algorithm
      • Orthogonal matrix and projector
      • Outer form of matrix-matrix product
      • Preconditioned Conjugate Gradients algorithm
      • Preconditioned Conjugate Gradients code
      • Preconditioning
      • Preconditioning the Conjugate Gradients algorithm
      • Projection
      • Pythagorean theorem
      • QR factorization
      • QR factorization and least-squares
      • QR iteration
      • QR iteration 2x2 example
      • QR iteration for upper Hessenberg matrices
      • QR iteration with shift
      • QR using Givens transformations
      • QR using Householder transformations
      • Rajat's painless Conjugate Gradients
      • Residuals and solution increments in CG
      • Row pivoting
      • Schur decomposition
      • Sensitivity analysis
      • Sherman-Morrison-Woodbury formula
      • Singular value decomposition
      • Solving linear systems
      • Solving linear systems using LU
      • Solving triangular systems
      • Some orthogonality relations in CG
      • SOR iteration
      • SOR iteration as a splitting method
      • Space and time costs of CG and GMRES
      • Splitting methods
      • Stability of the Cholesky factorization
      • Stability of the LU factorization
      • Subspace and linear independence
      • Summary of convergence and cost of the QR iteration
      • Summary of least-squares solution methods
      • Symmetric and unsymmetric QR iteration
      • Symmetric Positive Definite Matrices
      • The four fundamental spaces
      • Three-term recurrence
      • Trace
      • Triangular factorization
      • Uniqueness of the QR factorization
      • Unitarily diagonalizable matrices
      • Upper Hessenberg form for the QR iteration
      • Vector norms
      • Vectors and matrices
      • Why eigenvalues
    Home

    ❯

    Krylov iterative methods to solve sparse linear systems

    Krylov iterative methods to solve sparse linear systems

    Dec 05, 20245 min read

    • Krylov methods for sparse systems
      • This provides a basic introduction to all the methods covered in this chapter.
      • As alluded to previously, we search for the solution in the Krylov subspace that satisfies some appropriate optimality condition.
      • The optimality condition depends on the properties of the matrix: symmetric positive definite, symmetric, and general matrix.
    • Conjugate Gradients Version 1
      • This is a short and basic derivation of the algorithm.
      • This method produces the correct sequence of iterates xk​.
      • However, its space (memory) and time (number of flops) costs are not optimal.
      • This will be refined in the sections below.
    • Some orthogonality relations in CG
      • We illustrate the first of several orthogonality relations satisfied by the CG vector sequences.
      • The orthogonality relation is special. It is defined using a scaling by A1/2.
      • In this section, we prove that the solution increments Δx(k) are A-conjugate to each other (i.e., orthogonal with A1/2 scaling).
    • Convergence with conjugate steps
      • Based on the derivation from the previous section, we can prove that CV converges in at most n steps, where n is the size of A.
    • Residuals and solution increments in CG
      • We now start deriving a series of results to help us turn the current version of CG into an efficient algorithm.
      • We prove that the subspace spanned by the residuals is the same as the subspace spanned by the solution increments.
    • CG search directions
      • Using the previous section, we establish that the residuals can be written as linear combinations of the increments.
      • We will see later on how the coefficients in the linear combination can be easily computed using orthogonality relations.
    • All the orthogonality relations in CG
      • We derive all the orthogonality relations in CG that you need to know.
      • This is the foundation to build the final steps of CG in the sections below.
    • Three-term recurrence
      • Using the previous result and the orthogonality relations, we derive a three-term recurrence for the search directions p(l).
    • Optimal step size
      • Using our orthogonality relations, we derive a computationally efficient formula for the step sizes μk+1​.
    • Computationally efficient search directions
      • We need one more step to make computing the search directions p(k) efficient.
    • Conjugate Gradients algorithm
      • Based on all the previous equations, we describe each step in the CG algorithm.
    • Conjugate Gradients code
      • We give the Julia code for CG based on the algorithm we derived.
    • Convergence of the Conjugate Gradients
      • We give an estimate for the convergence of CG.
      • In some sense, this convergence can be considered to be optimal.
    • Rajat’s painless Conjugate Gradients
      • Rajat’s own derivation of the CG algorithm.
    • 2024 complete and painless Conjugate Gradient
      • Another standalone derivation of the CG algorithm based on Rajat’s derivation.
      • This derivation contains all the key results.
      • This is the recommended reading for understanding CG.
    • GMRES
      • CG applies when the matrix is symmetric positive definite.
      • GMRES applies to general unsymmetric matrices.
      • It requires a new definition of the norm used in the optimization problem.
    • GMRES least-squares problem
      • We derive the least-squares problem GMRES is solving.
    • GMRES algorithm
      • We explain the steps in the GMRES algorithm based on the previous derivation.
    • Convergence of GMRES
      • The convergence of GMRES depends on the distribution of the eigenvalues. This is similar to CG.
      • It also depends on the condition number of the eigenvector basis matrix X.
    • Space and time costs of CG and GMRES
      • We derive the space and time computational cost for CG and GMRES.
      • As expected, GMRES has a larger computational cost. GMRES has an extra factor k in the cost compared to CG.
    • MINRES
      • This algorithm specializes GMRES for symmetric matrices.
      • The space cost is reduced to O(n) and the time cost to O(k(nnz+n)).
    • Preconditioning
      • This technique is used to reduce the number of iterations in iterative methods.
      • It works by changing the way eigenvalues are distributed.
      • Preconditioners need to be computationally cheap to apply and should reduce the condition number of the linear system.
    • Preconditioning the Conjugate Gradients algorithm
      • We explain the basic idea behind preconditioning CG.
      • Although we are working with CACT, because the Krylov subspace is based on using powers of the matrix, we can replace C and CT by the simpler M=CTC matrix.
      • Then any preconditioner M that is SPD and such that MA≈I can be used to precondition CG.
      • The only condition we require is that M should be computationally cheap to apply.
    • Preconditioned Conjugate Gradients algorithm
      • We describe how our previous idea can be simplified to reduce the computational cost.
      • The matrices C and CT are replaced by M and disappear entirely from the algorithm.
      • The final algorithm is very close to the original CG algorithm.
    • Preconditioned Conjugate Gradients code
      • We provide a short Julia code implementing the PCG algorithm.
    • Flexible preconditioned conjugate gradient method
      • This preconditioning technique is a variant of PCG that allows using a preconditioner like M(k) that varies at every step.
      • This can be very useful in some circumstances.

    Graph View

    Backlinks

    • CME 302 Numerical Linear Algebra

    Created with Quartz v4.3.1 © 2024

    • GitHub
    • Canvas