×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide

Solutions for Chapter 2.1: Systems of Linear Equations: Substitution; Elimination

Full solutions for Finite Mathematics, Binder Ready Version: An Applied Approach | 11th Edition

ISBN: 9780470876398

Solutions for Chapter 2.1: Systems of Linear Equations: Substitution; Elimination

Solutions for Chapter 2.1
4 5 0 246 Reviews
15
4
Textbook: Finite Mathematics, Binder Ready Version: An Applied Approach
Edition: 11
Author: Michael Sullivan
ISBN: 9780470876398

This expansive textbook survival guide covers the following chapters and their solutions. Chapter 2.1: Systems of Linear Equations: Substitution; Elimination includes 70 full step-by-step solutions. This textbook survival guide was created for the textbook: Finite Mathematics, Binder Ready Version: An Applied Approach, edition: 11. Finite Mathematics, Binder Ready Version: An Applied Approach was written by and is associated to the ISBN: 9780470876398. Since 70 problems in chapter 2.1: Systems of Linear Equations: Substitution; Elimination have been answered, more than 16741 students have viewed full step-by-step solutions from this chapter.

Key Math Terms and definitions covered in this textbook
  • Affine transformation

    Tv = Av + Vo = linear transformation plus shift.

  • Associative Law (AB)C = A(BC).

    Parentheses can be removed to leave ABC.

  • Eigenvalue A and eigenvector x.

    Ax = AX with x#-O so det(A - AI) = o.

  • Fast Fourier Transform (FFT).

    A factorization of the Fourier matrix Fn into e = log2 n matrices Si times a permutation. Each Si needs only nl2 multiplications, so Fnx and Fn-1c can be computed with ne/2 multiplications. Revolutionary.

  • Four Fundamental Subspaces C (A), N (A), C (AT), N (AT).

    Use AT for complex A.

  • Free columns of A.

    Columns without pivots; these are combinations of earlier columns.

  • Kirchhoff's Laws.

    Current Law: net current (in minus out) is zero at each node. Voltage Law: Potential differences (voltage drops) add to zero around any closed loop.

  • Krylov subspace Kj(A, b).

    The subspace spanned by b, Ab, ... , Aj-Ib. Numerical methods approximate A -I b by x j with residual b - Ax j in this subspace. A good basis for K j requires only multiplication by A at each step.

  • Left inverse A+.

    If A has full column rank n, then A+ = (AT A)-I AT has A+ A = In.

  • Linear combination cv + d w or L C jV j.

    Vector addition and scalar multiplication.

  • Linear transformation T.

    Each vector V in the input space transforms to T (v) in the output space, and linearity requires T(cv + dw) = c T(v) + d T(w). Examples: Matrix multiplication A v, differentiation and integration in function space.

  • Lucas numbers

    Ln = 2,J, 3, 4, ... satisfy Ln = L n- l +Ln- 2 = A1 +A~, with AI, A2 = (1 ± -/5)/2 from the Fibonacci matrix U~]' Compare Lo = 2 with Fo = O.

  • Orthogonal matrix Q.

    Square matrix with orthonormal columns, so QT = Q-l. Preserves length and angles, IIQxll = IIxll and (QX)T(Qy) = xTy. AlllAI = 1, with orthogonal eigenvectors. Examples: Rotation, reflection, permutation.

  • Pascal matrix

    Ps = pascal(n) = the symmetric matrix with binomial entries (i1~;2). Ps = PL Pu all contain Pascal's triangle with det = 1 (see Pascal in the index).

  • Permutation matrix P.

    There are n! orders of 1, ... , n. The n! P 's have the rows of I in those orders. P A puts the rows of A in the same order. P is even or odd (det P = 1 or -1) based on the number of row exchanges to reach I.

  • Pseudoinverse A+ (Moore-Penrose inverse).

    The n by m matrix that "inverts" A from column space back to row space, with N(A+) = N(AT). A+ A and AA+ are the projection matrices onto the row space and column space. Rank(A +) = rank(A).

  • Row picture of Ax = b.

    Each equation gives a plane in Rn; the planes intersect at x.

  • Simplex method for linear programming.

    The minimum cost vector x * is found by moving from comer to lower cost comer along the edges of the feasible set (where the constraints Ax = b and x > 0 are satisfied). Minimum cost at a comer!

  • Skew-symmetric matrix K.

    The transpose is -K, since Kij = -Kji. Eigenvalues are pure imaginary, eigenvectors are orthogonal, eKt is an orthogonal matrix.

  • Subspace S of V.

    Any vector space inside V, including V and Z = {zero vector only}.

×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide
×
Reset your password