Exam 2 Study Guide/Practice Aid

Content

The exam covers content from LU factorization through Schur form. Roughly in terms of subsections, that is

  • LU factorization
  • LU: failure cases and pivoting
  • Cholesky factorization
  • LU: cost, BLAS/LAPACK, Schur complement, round-off error
  • Condition number via SVD
  • Least squares: setup and geometric interpretation
  • Least squares: normal equations and orthogonal projectors
  • Least squares: pseudoinverse
  • Least squares: sensitivity and conditioning
  • QR factorization: Householder reflectors and Givens rotations
  • SVD: reduced/full forms and applications
  • Eigenvalue problems: setup, characteristic polynomial, multiplicity
  • Similar matrices and diagonalizability
  • Eigenvalue transformations (shift, inversion, power, polynomial)

Some questions to ask yourself

LU Factorization

  • What is the LU factorization? What are the shapes and properties of $L$ and $U$?
  • Can I carry out LU factorization on a small matrix by hand?
  • What is the one failure mode of LU without pivoting? Given a matrix, can I determine whether LU without pivoting exists?
  • What is partial pivoting? What is complete pivoting? What are the trade-offs?
  • What does a pivoted LU factorization look like? (i.e. $PA = LU$)
  • Given an LU factorization $PA = LU$, can I outline the solve process for $A\boldsymbol x = \boldsymbol b$?
  • What is the computational cost of LU factorization? Of forward/backward substitution? Of solving multiple right-hand sides? In special cases, such as banded matrix?
  • What are the BLAS levels, and why do they matter for the constant in the $O(n^3)$ cost?
  • What is the Schur complement? How does block Gaussian elimination produce it? Can I rederive the formula if needed?
  • How does round-off error in LU behave with and without pivoting? Can I trace through the $2\times 2$ example with $\epsilon < \epsilon_\text{mach}$?

Cholesky Factorization

  • What is the Cholesky factorization? For what class of matrices does it apply?
  • Why does Cholesky require no pivoting to guarantee existence for SPD matrices?
  • What is the relationship (if any) between the LU and Cholesky factorizations of an SPD matrix?
  • How do the singular values of $A$ relate to the diagonal of $L$ in $A = LL^T$?
  • What is the cost advantage of Cholesky over LU?

Condition Number and SVD

  • What is the SVD $A = U\Sigma V^T$? What are the shapes and properties of the factors?
  • How does the 2-norm of a matrix relate to its singular values?
  • How is $\text{cond}_2(A)$ expressed in terms of the singular values?
  • Given the SVD of a matrix, can I compute its condition number?
  • What is the Frobenius norm? Is it an induced norm? How does it relate to the SVD?

Linear Least Squares: Setup

  • Can I set up the matrix for a function fitting problem? In one dimension? In multiple dimensions?
  • What is the least squares problem $A\boldsymbol x \cong \boldsymbol b$? When does it arise?
  • What are the normal equations? How are they derived?
  • What is the geometric interpretation of the least squares solution? (orthogonal projection, residual $\perp$ column space of $A$)
  • Given vectors $\boldsymbol b$, $\boldsymbol y = A\boldsymbol x$, and the angle $\theta$ between $\boldsymbol b$ and $\text{colspan}(A)$, can I interpret $\|\boldsymbol b\|_2 \cos\theta$?
  • What is an orthogonal projector? What conditions characterize one?
  • Can I verify that $P = A(A^TA)^{-1}A^T$ is an orthogonal projector onto $\text{colspan}(A)$?
  • What is the pseudoinverse $A^+$ of a full-column-rank matrix?

Least Squares: Sensitivity and Conditioning

  • What is the conditioning bound for the least squares solution when $\boldsymbol b$ is perturbed? [ \frac{|\Delta\boldsymbol x|_2}{|\boldsymbol x|_2} \leq \kappa(A) \frac{1}{\cos\theta} \frac{|\Delta\boldsymbol b|_2}{|\boldsymbol b|_2} ]
  • Can I apply this bound to compute a numerical answer?
  • Which values of $\theta$ make the least squares problem ill-conditioned? Why?
  • How does the sensitivity depend on both $A$ and $\boldsymbol b$ (unlike for square linear systems)?

QR Factorization

  • What is the QR factorization? How does it reduce a least squares problem to a triangular one?
  • How can a QR factorization be used to solve least squares problems?
  • How can Gram-Schmidt be used to calculate QR? What is modified Gram-Schmidt? How does Gram-Schmidt fall short numerically?
  • How does a Householder reflector $H = I - 2\frac{\boldsymbol v \boldsymbol v^T}{\boldsymbol v^T \boldsymbol v}$ zero out entries below the first in a vector?
  • Can I compute the first Householder reflector for a given matrix, and apply it to find $H_1 A$?
  • Which sign choice for $\boldsymbol v = \boldsymbol a \mp \|\boldsymbol a\|_2 \boldsymbol e_1$ minimizes cancellation?
  • What are Givens rotations? How do they produce zeros one at a time? In what order are they applied to a $3\times 3$ matrix?
  • What is economical/reduced QR? How does it factor into LSQ solutions?
  • What is rank-revealing QR (column pivoting)? What factorization does it produce?

SVD: Applications

  • What are the shapes of the factors in the full vs.\ economical SVD for an $m\times n$ matrix with $m > n$?
  • What does the SVD reveal about the nullspace, rank, and numerical rank of a matrix?
  • What is the Eckart-Young-Mirsky theorem? Given a truncated SVD $A_k$, what is $\|A - A_k\|_2$? What aobut the Frobenius norm?
  • How is the minimum-norm least squares solution expressed via the pseudoinverse $A^+ = V\Sigma^+ U^T$?
  • Can I compare the cost of solving least squares via normal equations (Cholesky), Householder QR, and SVD?

Eigenvalue Problems: Setup

  • What is an eigenvalue? An eigenvector? The spectrum? The spectral radius?
  • What is the characteristic polynomial? Why does it not yield a practical algorithm for $n \geq 5$? Why is it numerically not useful?
  • What are algebraic and geometric multiplicity? What does it mean for a matrix to be defective?
  • What does it mean for a matrix to be diagonalizable? Can I write $A = XDX^{-1}$ given the eigenvectors?

Similar Matrices

  • What does it mean for two matrices to be similar?
  • What properties are shared by similar matrices? (eigenvalues/spectral radius, diagonalizability, characteristic polynomial)
  • What properties are not necessarily shared? (singular values, symmetry, orthogonality of eigenvectors)
  • Can I verify whether a given statement about similar matrices must be true or false?

Eigenvalue Transformations

  • Given $A\boldsymbol x = \lambda \boldsymbol x$, what are the eigenvalues of:
    • $A - \sigma I$ (shift)
    • $A^{-1}$ (inversion)
    • $A^k$ (power)
    • $p(A)$ for a polynomial $p$ (polynomial)
    • $T^{-1}AT$ (similarity)
  • Can I compute a specific eigenvalue of a transformed matrix given an eigenvalue of the original?