×
Log in to StudySoup
Get Full Access to Math - Textbook Survival Guide
Join StudySoup for FREE
Get Full Access to Math - Textbook Survival Guide

Already have an account? Login here
×
Reset your password

Solutions for Chapter 10.4: Even and Odd Functions

Elementary Differential Equations and Boundary Value Problems | 10th Edition | ISBN: 9780470458310 | Authors: William E. Boyce

Full solutions for Elementary Differential Equations and Boundary Value Problems | 10th Edition

ISBN: 9780470458310

Elementary Differential Equations and Boundary Value Problems | 10th Edition | ISBN: 9780470458310 | Authors: William E. Boyce

Solutions for Chapter 10.4: Even and Odd Functions

Solutions for Chapter 10.4
4 5 0 251 Reviews
27
3
Textbook: Elementary Differential Equations and Boundary Value Problems
Edition: 10
Author: William E. Boyce
ISBN: 9780470458310

This textbook survival guide was created for the textbook: Elementary Differential Equations and Boundary Value Problems, edition: 10. This expansive textbook survival guide covers the following chapters and their solutions. Elementary Differential Equations and Boundary Value Problems was written by and is associated to the ISBN: 9780470458310. Chapter 10.4: Even and Odd Functions includes 40 full step-by-step solutions. Since 40 problems in chapter 10.4: Even and Odd Functions have been answered, more than 18249 students have viewed full step-by-step solutions from this chapter.

Key Math Terms and definitions covered in this textbook
  • Conjugate Gradient Method.

    A sequence of steps (end of Chapter 9) to solve positive definite Ax = b by minimizing !x T Ax - x Tb over growing Krylov subspaces.

  • Covariance matrix:E.

    When random variables Xi have mean = average value = 0, their covariances "'£ ij are the averages of XiX j. With means Xi, the matrix :E = mean of (x - x) (x - x) T is positive (semi)definite; :E is diagonal if the Xi are independent.

  • Dot product = Inner product x T y = XI Y 1 + ... + Xn Yn.

    Complex dot product is x T Y . Perpendicular vectors have x T y = O. (AB)ij = (row i of A)T(column j of B).

  • Eigenvalue A and eigenvector x.

    Ax = AX with x#-O so det(A - AI) = o.

  • Ellipse (or ellipsoid) x T Ax = 1.

    A must be positive definite; the axes of the ellipse are eigenvectors of A, with lengths 1/.JI. (For IIx II = 1 the vectors y = Ax lie on the ellipse IIA-1 yll2 = Y T(AAT)-1 Y = 1 displayed by eigshow; axis lengths ad

  • Gauss-Jordan method.

    Invert A by row operations on [A I] to reach [I A-I].

  • Graph G.

    Set of n nodes connected pairwise by m edges. A complete graph has all n(n - 1)/2 edges between nodes. A tree has only n - 1 edges and no closed loops.

  • Hermitian matrix A H = AT = A.

    Complex analog a j i = aU of a symmetric matrix.

  • Incidence matrix of a directed graph.

    The m by n edge-node incidence matrix has a row for each edge (node i to node j), with entries -1 and 1 in columns i and j .

  • Inverse matrix A-I.

    Square matrix with A-I A = I and AA-l = I. No inverse if det A = 0 and rank(A) < n and Ax = 0 for a nonzero vector x. The inverses of AB and AT are B-1 A-I and (A-I)T. Cofactor formula (A-l)ij = Cji! detA.

  • Linear transformation T.

    Each vector V in the input space transforms to T (v) in the output space, and linearity requires T(cv + dw) = c T(v) + d T(w). Examples: Matrix multiplication A v, differentiation and integration in function space.

  • Orthonormal vectors q 1 , ... , q n·

    Dot products are q T q j = 0 if i =1= j and q T q i = 1. The matrix Q with these orthonormal columns has Q T Q = I. If m = n then Q T = Q -1 and q 1 ' ... , q n is an orthonormal basis for Rn : every v = L (v T q j )q j •

  • Right inverse A+.

    If A has full row rank m, then A+ = AT(AAT)-l has AA+ = 1m.

  • Singular Value Decomposition

    (SVD) A = U:E VT = (orthogonal) ( diag)( orthogonal) First r columns of U and V are orthonormal bases of C (A) and C (AT), AVi = O'iUi with singular value O'i > O. Last columns are orthonormal bases of nullspaces.

  • Subspace S of V.

    Any vector space inside V, including V and Z = {zero vector only}.

  • Sum V + W of subs paces.

    Space of all (v in V) + (w in W). Direct sum: V n W = to}.

  • Triangle inequality II u + v II < II u II + II v II.

    For matrix norms II A + B II < II A II + II B II·

  • Unitary matrix UH = U T = U-I.

    Orthonormal columns (complex analog of Q).

  • Vector space V.

    Set of vectors such that all combinations cv + d w remain within V. Eight required rules are given in Section 3.1 for scalars c, d and vectors v, w.

  • Volume of box.

    The rows (or the columns) of A generate a box with volume I det(A) I.