Lecture 9: Solving the Normal Equations by QR and Gram-Schmidt
Summary
Discussed solution of normal equations. Discussed condition number of solving normal equations directly, and noted that it squares the condition number—not a good idea if we can avoid it.
Introduced the alternative of QR factorization (finding an orthonormal basis for the column space of the matrix). Explained why, if we can do it accurately, this will give a good way to solve least-squares problems.
Gave the simple, but unstable, construction of the Gram-Schmidt algorithm, to find a QR factorization. Analyzed its O(mn2) complexity (specifically, 2mn2 flops), and commented that the "same" projection qqᵀa requires O(m2) operations if you perform it as (qqT)a but O(m) operations if you perform it as q(qTa) — matrix operations are associative (but not commutative), but where you put the parentheses can make a big difference in performance!
Discussed loss of orthogonality in classical Gram-Schmidt, using a simple example, especially in the case where the matrix has nearly dependent columns to begin with.
Further Reading
- Read “Lectures 7, 8, 18, and 19” in the textbook Numerical Linear Algebra.
- Gram-Schmidt Orthogonalization (PDF) (Courtesy of Per-Olof Persson. Used with permission.)
- Gram-Schmidt process on Wikipedia.
Lecture 10: Modified Gram-Schmidt and Householder QR
Summary
Discussed loss of orthogonality in classical Gram-Schmidt, using a simple example, especially in the case where the matrix has nearly dependent columns to begin with. Showed modified Gram-Schmidt and argued how it (mostly) fixes the problem. Numerical examples (see notebook below).
Re-interpreted Gram-Schmidt in matrix form as Q = AR1R2..., i.e. as multiplying A on the right by a sequence of upper-triangular matrices to get Q. The problem is that these matrices R may be very badly conditioned, leading to an inaccurate Q and loss of orthogonality.
Instead of multiplying A on the right by R's to get Q, however, we can instead multiply A on the left by Q's to get R. This leads us to the Householder QR algorithm. Introduced Householder QR, emphasized the inherent stability properties of multiplying by a sequence of unitary matrices (as shown in Problem set 2). Showed how we can convert a matrix to upper-triangular form (superficially similar to Gaussian elimination) via unitary Householder reflectors.
- Lecture 10 handout: Householder Reflectors and Givens Rotations (PDF) (Courtesy of Per-Olof Persson. Used with permission.)
- Lecture 10 notebook: Classical vs. Modified Gram-Schmidt
Further Reading
- Read “Lectures 7, 8, 16, 18, and 19” in the textbook Numerical Linear Algebra. It turns out that modified GS is backwards stable in the sense that the product QR is close to A, i.e. the function f(A) = Q*R is backwards stable in MGS; this is why solving systems with Q, R (appropriately used as discussed in Trefethen Lecture 19) is an accurate approximation to solving them with A.
- For a review of the literature on backwards-stability proofs of MGS, see Paige, Christopher C., Miroslav Rozlozník, and Zdenvek Strakos "Modified Gram-Schmidt (MGS), Least Squares, and Backward Stability of MGS-GMRES." SIAM J. Matrix Anal. Appl. 28, pp. 264–284.
Lecture 11: Matrix Operations, Caches, and Blocking
Summary
Finished Householder QR derivation, including the detail that one has a choice of Householder reflectors...we choose the sign to avoid taking differences of nearly-equal vectors. Gave flop count, showed that we don't need to explicitly compute Q if we store the Householder reflector vectors.
Counting arithmetic operation counts is no longer enough. Illustrate this with some performance experiments on a much simpler problem, matrix multiplication (see handouts). This leads us to analyze memory-access efficiency and caches and points the way to restructuring many algorithms.
Outline of the memory hierarchy: CPU, registers, L1/L2 cache, main memory, and presented simple 2-level ideal-cache model that we can analyze to get the basic ideas.
Analyzed cache complexity of simple row-column matrix multiply, showed that it asymptotically gets no benefit from the cache. Presented blocked algorithm, and showed that it achieves optimal Θ(n³/√Z) cache complexity.
Discussed some practical difficulties of the blocked matrix multiplication: algorithm depends on cache-size Z, and multi-level memories require multi-level blocking. Discussed how these ideas are applied to the design of modern linear-algebra libraries (LAPACK) by building them out of block operations (performed by an optimized BLAS).
- Lecture 11 handout: Performance Experiments with Matrix Multiplication (PDF)
- Ideal-Cache Terminology (PDF)
Assignment
Further Reading
- Wikipedia has a reasonable introduction to "Locality of Reference" that you might find useful.
- The optimized matrix multiplication shown on the handouts is called "Automatically Tuned Linear Algebra Software (ATLAS)."
- "Cache-Oblivious Algorithms" describes ideal cache model and analysis for various algorithms.
- "MATLAB Incorporates LAPACK" is about the switch from LINPACK to LAPACK/BLAS in MATLAB.
- The lecture video "Cache-Efficient Algorithms" in 6.172 Performance Engineering of Software Systems include discussions of matrix multiplication.