×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide

Solutions for Chapter 8.3: Compositions and Inverse Transformations

Full solutions for Elementary Linear Algebra: Applications Version | 10th Edition

ISBN: 9780470432051

Solutions for Chapter 8.3: Compositions and Inverse Transformations

Solutions for Chapter 8.3
4 5 0 285 Reviews
16
1
Textbook: Elementary Linear Algebra: Applications Version
Edition: 10
Author: Howard Anton, Chris Rorres
ISBN: 9780470432051

Elementary Linear Algebra: Applications Version was written by and is associated to the ISBN: 9780470432051. Chapter 8.3: Compositions and Inverse Transformations includes 29 full step-by-step solutions. This textbook survival guide was created for the textbook: Elementary Linear Algebra: Applications Version, edition: 10. This expansive textbook survival guide covers the following chapters and their solutions. Since 29 problems in chapter 8.3: Compositions and Inverse Transformations have been answered, more than 13761 students have viewed full step-by-step solutions from this chapter.

Key Math Terms and definitions covered in this textbook
  • Associative Law (AB)C = A(BC).

    Parentheses can be removed to leave ABC.

  • Block matrix.

    A matrix can be partitioned into matrix blocks, by cuts between rows and/or between columns. Block multiplication ofAB is allowed if the block shapes permit.

  • Commuting matrices AB = BA.

    If diagonalizable, they share n eigenvectors.

  • Eigenvalue A and eigenvector x.

    Ax = AX with x#-O so det(A - AI) = o.

  • Free columns of A.

    Columns without pivots; these are combinations of earlier columns.

  • Full column rank r = n.

    Independent columns, N(A) = {O}, no free variables.

  • Hermitian matrix A H = AT = A.

    Complex analog a j i = aU of a symmetric matrix.

  • Inverse matrix A-I.

    Square matrix with A-I A = I and AA-l = I. No inverse if det A = 0 and rank(A) < n and Ax = 0 for a nonzero vector x. The inverses of AB and AT are B-1 A-I and (A-I)T. Cofactor formula (A-l)ij = Cji! detA.

  • Kronecker product (tensor product) A ® B.

    Blocks aij B, eigenvalues Ap(A)Aq(B).

  • Krylov subspace Kj(A, b).

    The subspace spanned by b, Ab, ... , Aj-Ib. Numerical methods approximate A -I b by x j with residual b - Ax j in this subspace. A good basis for K j requires only multiplication by A at each step.

  • Multiplier eij.

    The pivot row j is multiplied by eij and subtracted from row i to eliminate the i, j entry: eij = (entry to eliminate) / (jth pivot).

  • Orthogonal subspaces.

    Every v in V is orthogonal to every w in W.

  • Orthonormal vectors q 1 , ... , q n·

    Dot products are q T q j = 0 if i =1= j and q T q i = 1. The matrix Q with these orthonormal columns has Q T Q = I. If m = n then Q T = Q -1 and q 1 ' ... , q n is an orthonormal basis for Rn : every v = L (v T q j )q j •

  • Row space C (AT) = all combinations of rows of A.

    Column vectors by convention.

  • Schwarz inequality

    Iv·wl < IIvll IIwll.Then IvTAwl2 < (vT Av)(wT Aw) for pos def A.

  • Similar matrices A and B.

    Every B = M-I AM has the same eigenvalues as A.

  • Simplex method for linear programming.

    The minimum cost vector x * is found by moving from comer to lower cost comer along the edges of the feasible set (where the constraints Ax = b and x > 0 are satisfied). Minimum cost at a comer!

  • Standard basis for Rn.

    Columns of n by n identity matrix (written i ,j ,k in R3).

  • Sum V + W of subs paces.

    Space of all (v in V) + (w in W). Direct sum: V n W = to}.

  • Unitary matrix UH = U T = U-I.

    Orthonormal columns (complex analog of Q).

×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide
×
Reset your password