Solutions for Chapter 10.3: Matrix Representations of Graphs

Discrete Mathematics with Applications | 4th Edition | ISBN: 9780495391326 | Authors: Susanna S. Epp

Full solutions for Discrete Mathematics with Applications | 4th Edition

ISBN: 9780495391326

Discrete Mathematics with Applications | 4th Edition | ISBN: 9780495391326 | Authors: Susanna S. Epp

Solutions for Chapter 10.3: Matrix Representations of Graphs

Solutions for Chapter 10.3
4 5 0 381 Reviews
12
0
Textbook: Discrete Mathematics with Applications
Edition: 4th
Author: Susanna S. Epp
ISBN: 9780495391326

Discrete Mathematics with Applications was written by Sieva Kozinsky and is associated to the ISBN: 9780495391326. This textbook survival guide was created for the textbook: Discrete Mathematics with Applications , edition: 4th. Chapter 10.3: Matrix Representations of Graphs includes 23 full step-by-step solutions. This expansive textbook survival guide covers the following chapters and their solutions. Since 23 problems in chapter 10.3: Matrix Representations of Graphs have been answered, more than 24658 students have viewed full step-by-step solutions from this chapter.

Key Math Terms and definitions covered in this textbook
  • Condition number

    cond(A) = c(A) = IIAIlIIA-III = amaxlamin. In Ax = b, the relative change Ilox III Ilx II is less than cond(A) times the relative change Ilob III lib II· Condition numbers measure the sensitivity of the output to change in the input.

  • Covariance matrix:E.

    When random variables Xi have mean = average value = 0, their covariances "'£ ij are the averages of XiX j. With means Xi, the matrix :E = mean of (x - x) (x - x) T is positive (semi)definite; :E is diagonal if the Xi are independent.

  • Cramer's Rule for Ax = b.

    B j has b replacing column j of A; x j = det B j I det A

  • Full column rank r = n.

    Independent columns, N(A) = {O}, no free variables.

  • Full row rank r = m.

    Independent rows, at least one solution to Ax = b, column space is all of Rm. Full rank means full column rank or full row rank.

  • Gram-Schmidt orthogonalization A = QR.

    Independent columns in A, orthonormal columns in Q. Each column q j of Q is a combination of the first j columns of A (and conversely, so R is upper triangular). Convention: diag(R) > o.

  • Hankel matrix H.

    Constant along each antidiagonal; hij depends on i + j.

  • Hermitian matrix A H = AT = A.

    Complex analog a j i = aU of a symmetric matrix.

  • Hilbert matrix hilb(n).

    Entries HU = 1/(i + j -1) = Jd X i- 1 xj-1dx. Positive definite but extremely small Amin and large condition number: H is ill-conditioned.

  • Minimal polynomial of A.

    The lowest degree polynomial with meA) = zero matrix. This is peA) = det(A - AI) if no eigenvalues are repeated; always meA) divides peA).

  • Multiplication Ax

    = Xl (column 1) + ... + xn(column n) = combination of columns.

  • Orthonormal vectors q 1 , ... , q n·

    Dot products are q T q j = 0 if i =1= j and q T q i = 1. The matrix Q with these orthonormal columns has Q T Q = I. If m = n then Q T = Q -1 and q 1 ' ... , q n is an orthonormal basis for Rn : every v = L (v T q j )q j •

  • Partial pivoting.

    In each column, choose the largest available pivot to control roundoff; all multipliers have leij I < 1. See condition number.

  • Permutation matrix P.

    There are n! orders of 1, ... , n. The n! P 's have the rows of I in those orders. P A puts the rows of A in the same order. P is even or odd (det P = 1 or -1) based on the number of row exchanges to reach I.

  • Pivot columns of A.

    Columns that contain pivots after row reduction. These are not combinations of earlier columns. The pivot columns are a basis for the column space.

  • Positive definite matrix A.

    Symmetric matrix with positive eigenvalues and positive pivots. Definition: x T Ax > 0 unless x = O. Then A = LDLT with diag(D» O.

  • Right inverse A+.

    If A has full row rank m, then A+ = AT(AAT)-l has AA+ = 1m.

  • Rotation matrix

    R = [~ CS ] rotates the plane by () and R- 1 = RT rotates back by -(). Eigenvalues are eiO and e-iO , eigenvectors are (1, ±i). c, s = cos (), sin ().

  • Simplex method for linear programming.

    The minimum cost vector x * is found by moving from comer to lower cost comer along the edges of the feasible set (where the constraints Ax = b and x > 0 are satisfied). Minimum cost at a comer!

  • Trace of A

    = sum of diagonal entries = sum of eigenvalues of A. Tr AB = Tr BA.

×
Log in to StudySoup
Get Full Access to Thousands of Study Materials at Your School

Forgot password? Reset password here

Join StudySoup for FREE
Get Full Access to Thousands of Study Materials at Your School
Join with Email
Already have an account? Login here
Reset your password

I don't want to reset my password

Need help? Contact support

Need an Account? Is not associated with an account
Sign up
We're here to help

Having trouble accessing your account? Let us help you, contact support at +1(510) 944-1054 or support@studysoup.com

Got it, thanks!
Password Reset Request Sent An email has been sent to the email address associated to your account. Follow the link in the email to reset your password. If you're having trouble finding our email please check your spam folder
Got it, thanks!
Already have an Account? Is already in use
Log in
Incorrect Password The password used to log in with this account is incorrect
Try Again

Forgot password? Reset it here