×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide

Solutions for Chapter 14.1: Graphs, Paths, and Circuits

Thinking Mathematically | 6th Edition | ISBN: 9780321867322 | Authors: Robert F. Blitzer

Full solutions for Thinking Mathematically | 6th Edition

ISBN: 9780321867322

Thinking Mathematically | 6th Edition | ISBN: 9780321867322 | Authors: Robert F. Blitzer

Solutions for Chapter 14.1: Graphs, Paths, and Circuits

Solutions for Chapter 14.1
4 5 0 341 Reviews
13
4
Textbook: Thinking Mathematically
Edition: 6
Author: Robert F. Blitzer
ISBN: 9780321867322

This expansive textbook survival guide covers the following chapters and their solutions. Thinking Mathematically was written by and is associated to the ISBN: 9780321867322. This textbook survival guide was created for the textbook: Thinking Mathematically, edition: 6. Chapter 14.1: Graphs, Paths, and Circuits includes 74 full step-by-step solutions. Since 74 problems in chapter 14.1: Graphs, Paths, and Circuits have been answered, more than 62737 students have viewed full step-by-step solutions from this chapter.

Key Math Terms and definitions covered in this textbook
  • Augmented matrix [A b].

    Ax = b is solvable when b is in the column space of A; then [A b] has the same rank as A. Elimination on [A b] keeps equations correct.

  • Diagonal matrix D.

    dij = 0 if i #- j. Block-diagonal: zero outside square blocks Du.

  • Distributive Law

    A(B + C) = AB + AC. Add then multiply, or mUltiply then add.

  • Dot product = Inner product x T y = XI Y 1 + ... + Xn Yn.

    Complex dot product is x T Y . Perpendicular vectors have x T y = O. (AB)ij = (row i of A)T(column j of B).

  • Elimination.

    A sequence of row operations that reduces A to an upper triangular U or to the reduced form R = rref(A). Then A = LU with multipliers eO in L, or P A = L U with row exchanges in P, or E A = R with an invertible E.

  • Factorization

    A = L U. If elimination takes A to U without row exchanges, then the lower triangular L with multipliers eij (and eii = 1) brings U back to A.

  • Free columns of A.

    Columns without pivots; these are combinations of earlier columns.

  • Jordan form 1 = M- 1 AM.

    If A has s independent eigenvectors, its "generalized" eigenvector matrix M gives 1 = diag(lt, ... , 1s). The block his Akh +Nk where Nk has 1 's on diagonall. Each block has one eigenvalue Ak and one eigenvector.

  • Kirchhoff's Laws.

    Current Law: net current (in minus out) is zero at each node. Voltage Law: Potential differences (voltage drops) add to zero around any closed loop.

  • Linear transformation T.

    Each vector V in the input space transforms to T (v) in the output space, and linearity requires T(cv + dw) = c T(v) + d T(w). Examples: Matrix multiplication A v, differentiation and integration in function space.

  • Lucas numbers

    Ln = 2,J, 3, 4, ... satisfy Ln = L n- l +Ln- 2 = A1 +A~, with AI, A2 = (1 ± -/5)/2 from the Fibonacci matrix U~]' Compare Lo = 2 with Fo = O.

  • Nilpotent matrix N.

    Some power of N is the zero matrix, N k = o. The only eigenvalue is A = 0 (repeated n times). Examples: triangular matrices with zero diagonal.

  • Orthogonal subspaces.

    Every v in V is orthogonal to every w in W.

  • Partial pivoting.

    In each column, choose the largest available pivot to control roundoff; all multipliers have leij I < 1. See condition number.

  • Projection matrix P onto subspace S.

    Projection p = P b is the closest point to b in S, error e = b - Pb is perpendicularto S. p 2 = P = pT, eigenvalues are 1 or 0, eigenvectors are in S or S...L. If columns of A = basis for S then P = A (AT A) -1 AT.

  • Singular Value Decomposition

    (SVD) A = U:E VT = (orthogonal) ( diag)( orthogonal) First r columns of U and V are orthonormal bases of C (A) and C (AT), AVi = O'iUi with singular value O'i > O. Last columns are orthonormal bases of nullspaces.

  • Spectral Theorem A = QAQT.

    Real symmetric A has real A'S and orthonormal q's.

  • Unitary matrix UH = U T = U-I.

    Orthonormal columns (complex analog of Q).

  • Vandermonde matrix V.

    V c = b gives coefficients of p(x) = Co + ... + Cn_IXn- 1 with P(Xi) = bi. Vij = (Xi)j-I and det V = product of (Xk - Xi) for k > i.

  • Vector addition.

    v + w = (VI + WI, ... , Vn + Wn ) = diagonal of parallelogram.

×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide
×
Reset your password