×
Log in to StudySoup
Get Full Access to Algebra - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Algebra - Textbook Survival Guide

Already have an account? Login here
×
Reset your password

Solutions for Chapter 3.7: Graphing Linear Inequalities

Intermediate Algebra | 6th Edition | ISBN: 9780321785046 | Authors: Elayn El Martin-Gay

Full solutions for Intermediate Algebra | 6th Edition

ISBN: 9780321785046

Intermediate Algebra | 6th Edition | ISBN: 9780321785046 | Authors: Elayn El Martin-Gay

Solutions for Chapter 3.7: Graphing Linear Inequalities

Solutions for Chapter 3.7
4 5 0 259 Reviews
20
0
Textbook: Intermediate Algebra
Edition: 6
Author: Elayn El Martin-Gay
ISBN: 9780321785046

Since 70 problems in chapter 3.7: Graphing Linear Inequalities have been answered, more than 66825 students have viewed full step-by-step solutions from this chapter. Intermediate Algebra was written by and is associated to the ISBN: 9780321785046. This textbook survival guide was created for the textbook: Intermediate Algebra, edition: 6. Chapter 3.7: Graphing Linear Inequalities includes 70 full step-by-step solutions. This expansive textbook survival guide covers the following chapters and their solutions.

Key Math Terms and definitions covered in this textbook
  • Cayley-Hamilton Theorem.

    peA) = det(A - AI) has peA) = zero matrix.

  • Characteristic equation det(A - AI) = O.

    The n roots are the eigenvalues of A.

  • Condition number

    cond(A) = c(A) = IIAIlIIA-III = amaxlamin. In Ax = b, the relative change Ilox III Ilx II is less than cond(A) times the relative change Ilob III lib IIĀ· Condition numbers measure the sensitivity of the output to change in the input.

  • Dimension of vector space

    dim(V) = number of vectors in any basis for V.

  • Dot product = Inner product x T y = XI Y 1 + ... + Xn Yn.

    Complex dot product is x T Y . Perpendicular vectors have x T y = O. (AB)ij = (row i of A)T(column j of B).

  • Echelon matrix U.

    The first nonzero entry (the pivot) in each row comes in a later column than the pivot in the previous row. All zero rows come last.

  • Elimination matrix = Elementary matrix Eij.

    The identity matrix with an extra -eij in the i, j entry (i #- j). Then Eij A subtracts eij times row j of A from row i.

  • Factorization

    A = L U. If elimination takes A to U without row exchanges, then the lower triangular L with multipliers eij (and eii = 1) brings U back to A.

  • Fast Fourier Transform (FFT).

    A factorization of the Fourier matrix Fn into e = log2 n matrices Si times a permutation. Each Si needs only nl2 multiplications, so Fnx and Fn-1c can be computed with ne/2 multiplications. Revolutionary.

  • Free variable Xi.

    Column i has no pivot in elimination. We can give the n - r free variables any values, then Ax = b determines the r pivot variables (if solvable!).

  • Hermitian matrix A H = AT = A.

    Complex analog a j i = aU of a symmetric matrix.

  • Hilbert matrix hilb(n).

    Entries HU = 1/(i + j -1) = Jd X i- 1 xj-1dx. Positive definite but extremely small Amin and large condition number: H is ill-conditioned.

  • Identity matrix I (or In).

    Diagonal entries = 1, off-diagonal entries = 0.

  • Markov matrix M.

    All mij > 0 and each column sum is 1. Largest eigenvalue A = 1. If mij > 0, the columns of Mk approach the steady state eigenvector M s = s > O.

  • Projection matrix P onto subspace S.

    Projection p = P b is the closest point to b in S, error e = b - Pb is perpendicularto S. p 2 = P = pT, eigenvalues are 1 or 0, eigenvectors are in S or S...L. If columns of A = basis for S then P = A (AT A) -1 AT.

  • Projection p = a(aTblaTa) onto the line through a.

    P = aaT laTa has rank l.

  • Singular Value Decomposition

    (SVD) A = U:E VT = (orthogonal) ( diag)( orthogonal) First r columns of U and V are orthonormal bases of C (A) and C (AT), AVi = O'iUi with singular value O'i > O. Last columns are orthonormal bases of nullspaces.

  • Skew-symmetric matrix K.

    The transpose is -K, since Kij = -Kji. Eigenvalues are pure imaginary, eigenvectors are orthogonal, eKt is an orthogonal matrix.

  • Spectral Theorem A = QAQT.

    Real symmetric A has real A'S and orthonormal q's.

  • Standard basis for Rn.

    Columns of n by n identity matrix (written i ,j ,k in R3).