×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide

Solutions for Chapter 2.3: Discrete Mathematics with Applications 4th Edition

Discrete Mathematics with Applications | 4th Edition | ISBN: 9780495391326 | Authors: Susanna S. Epp

Full solutions for Discrete Mathematics with Applications | 4th Edition

ISBN: 9780495391326

Discrete Mathematics with Applications | 4th Edition | ISBN: 9780495391326 | Authors: Susanna S. Epp

Solutions for Chapter 2.3

Solutions for Chapter 2.3
4 5 0 272 Reviews
19
3
Textbook: Discrete Mathematics with Applications
Edition: 4
Author: Susanna S. Epp
ISBN: 9780495391326

Discrete Mathematics with Applications was written by and is associated to the ISBN: 9780495391326. This textbook survival guide was created for the textbook: Discrete Mathematics with Applications , edition: 4. Chapter 2.3 includes 44 full step-by-step solutions. Since 44 problems in chapter 2.3 have been answered, more than 47727 students have viewed full step-by-step solutions from this chapter. This expansive textbook survival guide covers the following chapters and their solutions.

Key Math Terms and definitions covered in this textbook
  • Back substitution.

    Upper triangular systems are solved in reverse order Xn to Xl.

  • Companion matrix.

    Put CI, ... ,Cn in row n and put n - 1 ones just above the main diagonal. Then det(A - AI) = ±(CI + c2A + C3A 2 + .•. + cnA n-l - An).

  • Cross product u xv in R3:

    Vector perpendicular to u and v, length Ilullllvlll sin el = area of parallelogram, u x v = "determinant" of [i j k; UI U2 U3; VI V2 V3].

  • Cyclic shift

    S. Permutation with S21 = 1, S32 = 1, ... , finally SIn = 1. Its eigenvalues are the nth roots e2lrik/n of 1; eigenvectors are columns of the Fourier matrix F.

  • Echelon matrix U.

    The first nonzero entry (the pivot) in each row comes in a later column than the pivot in the previous row. All zero rows come last.

  • Elimination matrix = Elementary matrix Eij.

    The identity matrix with an extra -eij in the i, j entry (i #- j). Then Eij A subtracts eij times row j of A from row i.

  • Full column rank r = n.

    Independent columns, N(A) = {O}, no free variables.

  • Gauss-Jordan method.

    Invert A by row operations on [A I] to reach [I A-I].

  • Gram-Schmidt orthogonalization A = QR.

    Independent columns in A, orthonormal columns in Q. Each column q j of Q is a combination of the first j columns of A (and conversely, so R is upper triangular). Convention: diag(R) > o.

  • Identity matrix I (or In).

    Diagonal entries = 1, off-diagonal entries = 0.

  • Kronecker product (tensor product) A ® B.

    Blocks aij B, eigenvalues Ap(A)Aq(B).

  • Norm

    IIA II. The ".e 2 norm" of A is the maximum ratio II Ax II/l1x II = O"max· Then II Ax II < IIAllllxll and IIABII < IIAIIIIBII and IIA + BII < IIAII + IIBII. Frobenius norm IIAII} = L La~. The.e 1 and.e oo norms are largest column and row sums of laij I.

  • Nullspace matrix N.

    The columns of N are the n - r special solutions to As = O.

  • Permutation matrix P.

    There are n! orders of 1, ... , n. The n! P 's have the rows of I in those orders. P A puts the rows of A in the same order. P is even or odd (det P = 1 or -1) based on the number of row exchanges to reach I.

  • Pivot columns of A.

    Columns that contain pivots after row reduction. These are not combinations of earlier columns. The pivot columns are a basis for the column space.

  • Row space C (AT) = all combinations of rows of A.

    Column vectors by convention.

  • Saddle point of I(x}, ... ,xn ).

    A point where the first derivatives of I are zero and the second derivative matrix (a2 II aXi ax j = Hessian matrix) is indefinite.

  • Simplex method for linear programming.

    The minimum cost vector x * is found by moving from comer to lower cost comer along the edges of the feasible set (where the constraints Ax = b and x > 0 are satisfied). Minimum cost at a comer!

  • Skew-symmetric matrix K.

    The transpose is -K, since Kij = -Kji. Eigenvalues are pure imaginary, eigenvectors are orthogonal, eKt is an orthogonal matrix.

  • Vector addition.

    v + w = (VI + WI, ... , Vn + Wn ) = diagonal of parallelogram.

×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide
×
Reset your password