- Thread starter
- #1

- Thread starter matqkks
- Start date

- Thread starter
- #1

- Feb 15, 2012

- 1,967

but, it can also be viewed the following ways:

it allows us to compute "the best possible orthogonal bases" of the domain and co-domain of a linear transformation of finite-dimensional linear spaces, in this sense that the matrix for T in these bases is as "simple as possible" (diagonal).

geometrically, this allows us to view any linear transformation as:

rotation+scaling map+rotation.

one way to see this is to "follow what happens to a unit n-sphere" (under the norm induced by the inner product we are using), for each of the three linear transformations in the decomposition.

it allows us to calculate the pseudo-inverse of a matrix, which is used in solving "least squares" (best fit) solutions such as finding the best fit polynomial of a given degree that matches the data (the polynomial isn't linear in its "indeterminate" variable, but IS a linear function of its coefficients).

in signal processing, the size of the singular values of a matrix are related to "which signals carry information" and "which signals are noise". calculating the SVD allows for "better (noise) filter design".

variations of the SVD are used in such diverse applications as: optical character recognition, radar target recognition profiles, 3d reconstruction from 2d images, fingerprint analysis, and weather prediction.

in general, calculation with a given mxn matrix is hard, evaluating the image of a given domain vector requires mn