Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide

Solutions for Chapter 1.2: Matrices

Full solutions for Elementary Linear Algebra with Applications | 9th Edition

ISBN: 9780471669593

Solutions for Chapter 1.2: Matrices

Solutions for Chapter 1.2
4 5 0 343 Reviews
Textbook: Elementary Linear Algebra with Applications
Edition: 9
Author: Howard Anton, Chris Rorres
ISBN: 9780471669593

Chapter 1.2: Matrices includes 24 full step-by-step solutions. Elementary Linear Algebra with Applications was written by and is associated to the ISBN: 9780471669593. This expansive textbook survival guide covers the following chapters and their solutions. Since 24 problems in chapter 1.2: Matrices have been answered, more than 9117 students have viewed full step-by-step solutions from this chapter. This textbook survival guide was created for the textbook: Elementary Linear Algebra with Applications, edition: 9.

Key Math Terms and definitions covered in this textbook
  • Change of basis matrix M.

    The old basis vectors v j are combinations L mij Wi of the new basis vectors. The coordinates of CI VI + ... + cnvn = dl wI + ... + dn Wn are related by d = M c. (For n = 2 set VI = mll WI +m21 W2, V2 = m12WI +m22w2.)

  • Diagonal matrix D.

    dij = 0 if i #- j. Block-diagonal: zero outside square blocks Du.

  • Diagonalizable matrix A.

    Must have n independent eigenvectors (in the columns of S; automatic with n different eigenvalues). Then S-I AS = A = eigenvalue matrix.

  • Dimension of vector space

    dim(V) = number of vectors in any basis for V.

  • Gauss-Jordan method.

    Invert A by row operations on [A I] to reach [I A-I].

  • Indefinite matrix.

    A symmetric matrix with eigenvalues of both signs (+ and - ).

  • Inverse matrix A-I.

    Square matrix with A-I A = I and AA-l = I. No inverse if det A = 0 and rank(A) < n and Ax = 0 for a nonzero vector x. The inverses of AB and AT are B-1 A-I and (A-I)T. Cofactor formula (A-l)ij = Cji! detA.

  • Markov matrix M.

    All mij > 0 and each column sum is 1. Largest eigenvalue A = 1. If mij > 0, the columns of Mk approach the steady state eigenvector M s = s > O.

  • Multiplication Ax

    = Xl (column 1) + ... + xn(column n) = combination of columns.

  • Norm

    IIA II. The ".e 2 norm" of A is the maximum ratio II Ax II/l1x II = O"max· Then II Ax II < IIAllllxll and IIABII < IIAIIIIBII and IIA + BII < IIAII + IIBII. Frobenius norm IIAII} = L La~. The.e 1 and.e oo norms are largest column and row sums of laij I.

  • Normal equation AT Ax = ATb.

    Gives the least squares solution to Ax = b if A has full rank n (independent columns). The equation says that (columns of A)·(b - Ax) = o.

  • Outer product uv T

    = column times row = rank one matrix.

  • Particular solution x p.

    Any solution to Ax = b; often x p has free variables = o.

  • Rayleigh quotient q (x) = X T Ax I x T x for symmetric A: Amin < q (x) < Amax.

    Those extremes are reached at the eigenvectors x for Amin(A) and Amax(A).

  • Reduced row echelon form R = rref(A).

    Pivots = 1; zeros above and below pivots; the r nonzero rows of R give a basis for the row space of A.

  • Semidefinite matrix A.

    (Positive) semidefinite: all x T Ax > 0, all A > 0; A = any RT R.

  • Special solutions to As = O.

    One free variable is Si = 1, other free variables = o.

  • Standard basis for Rn.

    Columns of n by n identity matrix (written i ,j ,k in R3).

  • Vandermonde matrix V.

    V c = b gives coefficients of p(x) = Co + ... + Cn_IXn- 1 with P(Xi) = bi. Vij = (Xi)j-I and det V = product of (Xk - Xi) for k > i.

  • Wavelets Wjk(t).

    Stretch and shift the time axis to create Wjk(t) = woo(2j t - k).

Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide
Reset your password