×
Log in to StudySoup
Get Full Access to Algebra - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Algebra - Textbook Survival Guide

Solutions for Chapter 3.3: Graphing Linear Functions

Intermediate Algebra | 6th Edition | ISBN: 9780321785046 | Authors: Elayn El Martin-Gay

Full solutions for Intermediate Algebra | 6th Edition

ISBN: 9780321785046

Intermediate Algebra | 6th Edition | ISBN: 9780321785046 | Authors: Elayn El Martin-Gay

Solutions for Chapter 3.3: Graphing Linear Functions

Solutions for Chapter 3.3
4 5 0 411 Reviews
30
0
Textbook: Intermediate Algebra
Edition: 6
Author: Elayn El Martin-Gay
ISBN: 9780321785046

Since 92 problems in chapter 3.3: Graphing Linear Functions have been answered, more than 58813 students have viewed full step-by-step solutions from this chapter. This expansive textbook survival guide covers the following chapters and their solutions. This textbook survival guide was created for the textbook: Intermediate Algebra, edition: 6. Intermediate Algebra was written by and is associated to the ISBN: 9780321785046. Chapter 3.3: Graphing Linear Functions includes 92 full step-by-step solutions.

Key Math Terms and definitions covered in this textbook
  • Affine transformation

    Tv = Av + Vo = linear transformation plus shift.

  • Change of basis matrix M.

    The old basis vectors v j are combinations L mij Wi of the new basis vectors. The coordinates of CI VI + ... + cnvn = dl wI + ... + dn Wn are related by d = M c. (For n = 2 set VI = mll WI +m21 W2, V2 = m12WI +m22w2.)

  • Column space C (A) =

    space of all combinations of the columns of A.

  • Complete solution x = x p + Xn to Ax = b.

    (Particular x p) + (x n in nullspace).

  • Conjugate Gradient Method.

    A sequence of steps (end of Chapter 9) to solve positive definite Ax = b by minimizing !x T Ax - x Tb over growing Krylov subspaces.

  • Cramer's Rule for Ax = b.

    B j has b replacing column j of A; x j = det B j I det A

  • Elimination matrix = Elementary matrix Eij.

    The identity matrix with an extra -eij in the i, j entry (i #- j). Then Eij A subtracts eij times row j of A from row i.

  • Gram-Schmidt orthogonalization A = QR.

    Independent columns in A, orthonormal columns in Q. Each column q j of Q is a combination of the first j columns of A (and conversely, so R is upper triangular). Convention: diag(R) > o.

  • Hermitian matrix A H = AT = A.

    Complex analog a j i = aU of a symmetric matrix.

  • Independent vectors VI, .. " vk.

    No combination cl VI + ... + qVk = zero vector unless all ci = O. If the v's are the columns of A, the only solution to Ax = 0 is x = o.

  • Inverse matrix A-I.

    Square matrix with A-I A = I and AA-l = I. No inverse if det A = 0 and rank(A) < n and Ax = 0 for a nonzero vector x. The inverses of AB and AT are B-1 A-I and (A-I)T. Cofactor formula (A-l)ij = Cji! detA.

  • Jordan form 1 = M- 1 AM.

    If A has s independent eigenvectors, its "generalized" eigenvector matrix M gives 1 = diag(lt, ... , 1s). The block his Akh +Nk where Nk has 1 's on diagonall. Each block has one eigenvalue Ak and one eigenvector.

  • Kronecker product (tensor product) A ® B.

    Blocks aij B, eigenvalues Ap(A)Aq(B).

  • Normal equation AT Ax = ATb.

    Gives the least squares solution to Ax = b if A has full rank n (independent columns). The equation says that (columns of A)·(b - Ax) = o.

  • Positive definite matrix A.

    Symmetric matrix with positive eigenvalues and positive pivots. Definition: x T Ax > 0 unless x = O. Then A = LDLT with diag(D» O.

  • Right inverse A+.

    If A has full row rank m, then A+ = AT(AAT)-l has AA+ = 1m.

  • Similar matrices A and B.

    Every B = M-I AM has the same eigenvalues as A.

  • Simplex method for linear programming.

    The minimum cost vector x * is found by moving from comer to lower cost comer along the edges of the feasible set (where the constraints Ax = b and x > 0 are satisfied). Minimum cost at a comer!

  • Singular Value Decomposition

    (SVD) A = U:E VT = (orthogonal) ( diag)( orthogonal) First r columns of U and V are orthonormal bases of C (A) and C (AT), AVi = O'iUi with singular value O'i > O. Last columns are orthonormal bases of nullspaces.

  • Vandermonde matrix V.

    V c = b gives coefficients of p(x) = Co + ... + Cn_IXn- 1 with P(Xi) = bi. Vij = (Xi)j-I and det V = product of (Xk - Xi) for k > i.

×
Log in to StudySoup
Get Full Access to Algebra - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Algebra - Textbook Survival Guide
×
Reset your password