×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide

Solutions for Chapter 57: Linear Extensions

Full solutions for Mathematics: A Discrete Introduction | 3rd Edition

ISBN: 9780840049421

Solutions for Chapter 57: Linear Extensions

Since 6 problems in chapter 57: Linear Extensions have been answered, more than 8858 students have viewed full step-by-step solutions from this chapter. This expansive textbook survival guide covers the following chapters and their solutions. Chapter 57: Linear Extensions includes 6 full step-by-step solutions. This textbook survival guide was created for the textbook: Mathematics: A Discrete Introduction, edition: 3. Mathematics: A Discrete Introduction was written by and is associated to the ISBN: 9780840049421.

Key Math Terms and definitions covered in this textbook
  • Complex conjugate

    z = a - ib for any complex number z = a + ib. Then zz = Iz12.

  • Dimension of vector space

    dim(V) = number of vectors in any basis for V.

  • Eigenvalue A and eigenvector x.

    Ax = AX with x#-O so det(A - AI) = o.

  • Hankel matrix H.

    Constant along each antidiagonal; hij depends on i + j.

  • Incidence matrix of a directed graph.

    The m by n edge-node incidence matrix has a row for each edge (node i to node j), with entries -1 and 1 in columns i and j .

  • Independent vectors VI, .. " vk.

    No combination cl VI + ... + qVk = zero vector unless all ci = O. If the v's are the columns of A, the only solution to Ax = 0 is x = o.

  • Iterative method.

    A sequence of steps intended to approach the desired solution.

  • Left inverse A+.

    If A has full column rank n, then A+ = (AT A)-I AT has A+ A = In.

  • Multiplier eij.

    The pivot row j is multiplied by eij and subtracted from row i to eliminate the i, j entry: eij = (entry to eliminate) / (jth pivot).

  • Network.

    A directed graph that has constants Cl, ... , Cm associated with the edges.

  • Orthogonal subspaces.

    Every v in V is orthogonal to every w in W.

  • Orthonormal vectors q 1 , ... , q n·

    Dot products are q T q j = 0 if i =1= j and q T q i = 1. The matrix Q with these orthonormal columns has Q T Q = I. If m = n then Q T = Q -1 and q 1 ' ... , q n is an orthonormal basis for Rn : every v = L (v T q j )q j •

  • Projection matrix P onto subspace S.

    Projection p = P b is the closest point to b in S, error e = b - Pb is perpendicularto S. p 2 = P = pT, eigenvalues are 1 or 0, eigenvectors are in S or S...L. If columns of A = basis for S then P = A (AT A) -1 AT.

  • Reflection matrix (Householder) Q = I -2uuT.

    Unit vector u is reflected to Qu = -u. All x intheplanemirroruTx = o have Qx = x. Notice QT = Q-1 = Q.

  • Simplex method for linear programming.

    The minimum cost vector x * is found by moving from comer to lower cost comer along the edges of the feasible set (where the constraints Ax = b and x > 0 are satisfied). Minimum cost at a comer!

  • Spectral Theorem A = QAQT.

    Real symmetric A has real A'S and orthonormal q's.

  • Spectrum of A = the set of eigenvalues {A I, ... , An}.

    Spectral radius = max of IAi I.

  • Symmetric matrix A.

    The transpose is AT = A, and aU = a ji. A-I is also symmetric.

  • Toeplitz matrix.

    Constant down each diagonal = time-invariant (shift-invariant) filter.

  • Trace of A

    = sum of diagonal entries = sum of eigenvalues of A. Tr AB = Tr BA.

×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide
×
Reset your password