Polar decomposition

Polar decomposition

In mathematics, particularly in linear algebra and functional analysis, the polar decomposition of a matrix or linear operator is a factorization analogous to the polar form of a nonzero complex number z

z = r e^{i \theta}\,

where r is the absolute value of z (a positive real number), and eiθ is called the complex sign of z.

Contents

Quaternion polar decomposition

The polar decomposition of quaternions H depends on the sphere \lbrace x i + y j + z k \in H : x^2 + y^2 +z^2 = 1 \rbrace of square roots of minus one. Given any r on this sphere, and an angle –π < a ≤ π, the versor e^{ar} = \cos (a) + r\ \sin (a) is on the 3-sphere of H. For a = 0 and a = π , the versor is 1 or −1 regardless of which r is selected. The norm t of a quaternion q is the Euclidean distance from the origin to q. When a quaternion is not just a real number, then there is a unique polar decomposition q = t e^{ar} \! .

Matrix polar decomposition

The polar decomposition of a square complex matrix A is a matrix decomposition of the form

A = UP\,

where U is a unitary matrix and P is a positive-semidefinite Hermitian matrix. Intuitively, the polar decomposition separates A into a component that stretches the space along a set of orthogonal axes, represented by P, and a rotation represented by U. The decomposition of the complex conjugate of A is given by \overline{A} = \overline{U} \overline{P}.

This decomposition always exists; and so long as A is invertible, it is unique, with P positive-definite. Note that

\det A = \det P\,\det U = re^{i\theta}

gives the corresponding polar decomposition of the determinant of A, since det P = r = | det A | and det U = eiθ.

The matrix P is always unique, even if A is singular, and given by

P = \sqrt{A^*A}

where A* denotes the conjugate transpose of A. This expression is meaningful since a positive-semidefinite Hermitian matrix has a unique positive-semidefinite square root. If A is invertible, then the matrix U is given by

U = AP^{-1}.\,

In terms of the singular value decomposition of A, A = W Σ V*, one has

P = V \Sigma V^*\,
U = W V^*\,

confirming that P is positive-definite and U is unitary.

One can also decompose A in the form

A = P'U\,

Here U is the same as before and P′ is given by

P' = UPU^{-1} = \sqrt{AA^*} = W \Sigma W^*.

This is known as the left polar decomposition, whereas the previous decomposition is known as the right polar decomposition. Left polar decomposition is also known as reverse polar decomposition.

The matrix A is normal if and only if P′ = P. Then UΣ = ΣU, and it is possible to diagonalise U with a unitary similarity matrix S that commutes with Σ, giving S U S* = Φ−1, where Φ is a diagonal unitary matrix of phases e. Putting Q = V S*, one can then re-write the polar decomposition as

 A = (Q \Phi Q^*)(Q \Sigma Q^*),\,

so A then thus also has a spectral decomposition

 A = Q  \Lambda Q^* \,

with complex eigenvalues such that ΛΛ* = Σ2 and a unitary matrix of complex eigenvectors Q.

The map from the general linear group GL(n,C) to the unitary group U(n) defined by mapping A onto its unitary piece U gives rise to a homotopy equivalence since the space of positive-definite matrices is contractible. In fact U(n) is a maximal compact subgroup of GL(n,C).

Bounded operators on Hilbert space

The polar decomposition of any bounded linear operator A between complex Hilbert spaces is a canonical factorization as the product of a partial isometry and a non-negative operator.

The polar decomposition for matrices generalizes as follows: if A is a bounded linear operator then there is a unique factorization of A as a product A = UP where U is a partial isometry, P is a non-negative self-adjoint operator and the initial space of U is the closure of the range of P.

The operator U must be weakened to a partial isometry, rather than unitary, because of the following issues. If A is the one-sided shift on l2(N), then |A| = {A*A}½ = I. So if A = U |A|, U must be A, which is not unitary.

The existence of a polar decomposition is a consequence of Douglas' lemma:

Lemma If A, B are bounded operators on a Hilbert space H, and A*AB*B, then there exists a contraction C such that A = CB. Furthermore, C is unique if Ker(B*) ⊂ Ker(C).

The operator C can be defined by C(Bh) = Ah, extended by continuity to the closure of Ran(B), and by zero on the orthogonal complement to all of H. The lemma then follows since A*AB*B implies Ker(A) ⊂ Ker(B).

In particular. If A*A = B*B, then C is a partial isometry, which is unique if Ker(B*) ⊂ Ker(C). In general, for any bounded operator A,

A^*A = (A^*A)^{\frac{1}{2}} (A^*A)^{\frac{1}{2}},

where (A*A)½ is the unique positive square root of A*A given by the usual functional calculus. So by the lemma, we have

A = U (A^*A)^{\frac{1}{2}}

for some partial isometry U, which is unique if Ker(A*) ⊂ Ker(U). Take P to be (A*A)½ and one obtains the polar decomposition A = UP. Notice that an analogous argument can be used to show A = P'U' , where P' is positive and U' a partial isometry.

When H is finite dimensional, U can be extended to a unitary operator; this is not true in general (see example above). Alternatively, the polar decomposition can be shown using the operator version of singular value decomposition.

By property of the continuous functional calculus, |A| is in the C*-algebra generated by A. A similar but weaker statement holds for the partial isometry: U is in the von Neumann algebra generated by A. If A is invertible, the polar part U will be in the C*-algebra as well.

Unbounded operators

If A is a closed, densely defined unbounded operator between complex Hilbert spaces then it still has a (unique) polar decomposition

A = U |A|\,

where |A| is a (possibly unbounded) non-negative self adjoint operator with the same domain as A, and U is a partial isometry vanishing on the orthogonal complement of the range Ran(|A|).

The proof uses the same lemma as above, which goes through for unbounded operators in general. If Dom(A*A) = Dom(B*B) and A*Ah = B*Bh for all hDom(A*A), then there exists a partial isometry U such that A = UB. U is unique if Ran(B)Ker(U). The operator A being closed and densely defined ensures that the operator A*A is self-adjoint (with dense domain) and therefore allows one to define (A*A)½. Applying the lemma gives polar decomposition.

If an unbounded operator A is affiliated to a von Neumann algebra M, and A = UP is its polar decomposition, then U is in M and so is the spectral projection of P, 1B(P), for any Borel set B in [0, ∞).

Alternative planar decompositions

In the Cartesian plane, alternative planar ring decompositions arise as follows:

  • If x ≠ 0, z = x ( 1 + (y/x) ε) is a polar decomposition of a dual number z = x + y ε, where ε ε = 0. In this polar decomposition, the unit circle has been replaced by the line x = 1, the polar angle by the slope y/x, and the radius x is negative in the left half-plane.
  • If x2 ≠ y2, then the unit hyperbola x2 − y2 = 1 and its conjugate x2 − y2 = −1 can be used to form a polar decomposition based on the branch of the unit hyperbola through (1,0). This branch is parametrized by the hyperbolic angle a and is written
\cosh (a)  + j \ \sinh (a)  = \exp(a j) = e^{a j}
where j j = +1 and the arithmetic of split-complex numbers is used. The branch through (−1,0) is traced by −ea j. Since the operation of multiplying by j reflects a point across the line y = x, the second hyperbola has branches traced by jea j or −jea j. Therefore a point in one of the quadrants has a polar decomposition in one of the forms:
r e^{a j}, - r e^ {a j}, r j e^{a j}, - r j e^{a j}, \quad r > 0
The set (1,−1,j,−j) has products that make it isomorphic to the Klein four-group. Evidently polar decomposition in this case involves an element from that group.

See also

References

  • Conway, J.B.: A Course in Functional Analysis. Graduate Texts in Mathematics. New York: Springer 1990
  • Douglas, R.G.: On Majorization, Factorization, and Range Inclusion of Operators on Hilbert Space. Proc. Amer. Math. Soc. 17, 413-415 (1966)

Wikimedia Foundation. 2010.

Игры ⚽ Нужно сделать НИР?

Look at other dictionaries:

  • polar ecosystem — Introduction       complex of living organisms in polar regions such as polar barrens and tundra.  Polar barrens and tundra are found at high latitudes on land surfaces not covered by perpetual ice and snow. These areas lying beyond the tree line …   Universalium

  • Singular value decomposition — Visualization of the SVD of a 2 dimensional, real shearing matrix M. First, we see the unit disc in blue together with the two canonical unit vectors. We then see the action of M, which distorts the disk to an ellipse. The SVD decomposes M into… …   Wikipedia

  • Cartan decomposition — The Cartan decomposition is a decomposition of a semisimple Lie group or Lie algebra, which plays an important role in their structure theory and representation theory. It generalizes the polar decomposition of matrices. Cartan involutions on Lie …   Wikipedia

  • Matrix decomposition — In the mathematical discipline of linear algebra, a matrix decomposition is a factorization of a matrix into some canonical form. There are many different matrix decompositions; each finds use among a particular class of problems. Contents 1… …   Wikipedia

  • QR decomposition — In linear algebra, the QR decomposition (also called the QR factorization) of a matrix is a decomposition of the matrix into an orthogonal and a right triangular matrix. The QR decomposition is often used to solve the linear least squares problem …   Wikipedia

  • Lie group decomposition — In mathematics, Lie group decompositions are used to analyse the structure of Lie groups and associated objects, by showing how they are built up out of subgroups. They are essential technical tools in the representation theory of Lie groups and… …   Wikipedia

  • Vector decomposition — refers to decomposing a vector of Rn into several vectors, each linearly independent (in mutually distinct directions in the n dimensional space). Vector decomposition in two dimensions In two dimensions, a vector can be decomposed in many ways.… …   Wikipedia

  • Orthogonal matrix — In linear algebra, an orthogonal matrix (less commonly called orthonormal matrix[1]), is a square matrix with real entries whose columns and rows are orthogonal unit vectors (i.e., orthonormal vectors). Equivalently, a matrix Q is orthogonal if… …   Wikipedia

  • Rotation matrix — In linear algebra, a rotation matrix is a matrix that is used to perform a rotation in Euclidean space. For example the matrix rotates points in the xy Cartesian plane counterclockwise through an angle θ about the origin of the Cartesian… …   Wikipedia

  • List of mathematics articles (P) — NOTOC P P = NP problem P adic analysis P adic number P adic order P compact group P group P² irreducible P Laplacian P matrix P rep P value P vector P y method Pacific Journal of Mathematics Package merge algorithm Packed storage matrix Packing… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”