Logarithm of a matrix

Logarithm of a matrix

In mathematics, a logarithm of a matrix is another matrix such that the matrix exponential of the latter matrix equals the original matrix. It is thus a generalization of the scalar logarithm and in some sense an inverse function of the matrix exponential. Not all matrices have a logarithm and those matrices that do have a logarithm may have more than one logarithm. The study of logarithms of matrices leads to Lie theory since when a matrix has a logarithm then it is in a Lie group and the logarithm is the corresponding element of the Lie algebra.

Contents

Definition

A matrix B is a logarithm of a given matrix A if the matrix exponential of B is A:

 e^B = A. \,

Example: Logarithm of rotations in the plane

The rotations in the plane give a simple example. A rotation of angle α around the origin is represented by the 2×2-matrix

 A =
\begin{pmatrix}
\cos(\alpha) & -\sin(\alpha) \\
\sin(\alpha) & \cos(\alpha) \\
\end{pmatrix}.

For any integer n, the matrix


B_n=(\alpha+2\pi n)
\begin{pmatrix}
0 & -1 \\
1 & 0\\
\end{pmatrix},

is a logarithm of A. Thus, the matrix A has infinitely many logarithms. This corresponds to the fact that the rotation angle is only determined up to multiples of 2π.

In the language of Lie theory, the rotation matrices A are elements of the Lie group SO(2). The corresponding logarithms B are elements of the Lie algebra so(2), which consists of all skew-symmetric matrices. The matrix


\begin{pmatrix}
0 & 1 \\
-1 & 0\\
\end{pmatrix}

is a generator of the Lie algebra so(2).

Existence

The question of whether a matrix has a logarithm has the easiest answer when considered in the complex setting. A matrix has a logarithm if and only if it is invertible.[1] The logarithm is not unique, but if a matrix has no negative real eigenvalues, then it has a unique logarithm whose eigenvalues lie all in the strip {zC | −π < Im z < π}. This logarithm is known as the principal logarithm.[2]

The answer is more involved in the real setting. A real matrix has a real logarithm if and only if it is invertible and each Jordan block belonging to a negative eigenvalue occurs an even number of times.[3] If an invertible real matrix does not satisfy the condition with the Jordan blocks, then it has only complex logarithms. This can already be seen in the scalar case: the logarithm of −1 is a complex number. The existence of real matrix logarithms of real 2 x 2 matrices is considered in a later section.

Properties

If A and B are both positive-definite matrices and A and B commute, i.e., AB = BA, then

AB = e^{\ln(A)+\ln(B)}. \,

For any invertible matrix,

A^{-1} = e^{-\ln(A)}. \,

Further example: Logarithm of rotations in 3D space

A rotation  R\in {\rm SO(3)} in the 3D is given by a 3x3 orthogonal matrix

The logarithm of the rotation matrix R can be computed with Rodrigues' rotation formula (see also Axis angle):

 \ln (R) = \left\{ \begin{matrix}
0 & \mathrm{if} \; \theta = 0 \\
\frac{\theta}{2 \sin(\theta)} (R - R^\top) & \mathrm{if} \; \theta \ne 0 \; \mathrm{and} \; \theta \in (-\pi, \pi)
  \end{matrix}\right.

Except when R has eigenvalues equal to −1 where the log is not unique. However, even in the case where θ = π the Frobenius norm of the log is:

 \| \ln(R) \|_F = \sqrt{2} | \theta |

Note that, given rotation matrices A and B:

 d_g(A,B) := \| \log(A^\top B)\|_F

is the geodesic distance on the 3D manifold of rotation matrices.

Calculating the logarithm of a diagonalizable matrix

A method for finding ln A for a diagonalizable matrix A is the following:

Find the matrix V of eigenvectors of A (each column of V is an eigenvector of A).
Find the inverse V−1 of V.
Let
 A' = V^{-1}  A  V.\,
Then A′ will be a diagonal matrix whose diagonal elements are eigenvalues of A.
Replace each diagonal element of A′ by its (natural) logarithm in order to obtain ln A'.
Then
 \ln A = V ( \ln A' ) V^{-1}. \,

That the logarithm of A might be a complex matrix even if A is real then follows from the fact that a matrix with real and positive entries might nevertheless have negative or even complex eigenvalues (this is true for example for rotation matrices). The non-uniqueness of the logarithm of a matrix follows from the non-uniqueness of the logarithm of a complex number.

The logarithm of a non-diagonalizable matrix

The algorithm illustrated above does not work for non-diagonalizable matrices, such as

\begin{bmatrix}1 & 1\\ 0 & 1\end{bmatrix}.

For such matrices one needs to find its Jordan decomposition and, rather than computing the logarithm of diagonal entries as above, one would calculate the logarithm of the Jordan blocks.

The latter is accomplished by noticing that one can write a Jordan block as

B=\begin{pmatrix}
\lambda & 1       & 0       & 0      & \cdots  & 0 \\
0       & \lambda & 1       & 0      & \cdots  & 0 \\
0       & 0       & \lambda & 1      & \cdots  & 0 \\
\vdots  & \vdots  & \vdots  & \ddots & \ddots  & \vdots \\
0       & 0       & 0       & 0      & \lambda & 1       \\
0       & 0       & 0       & 0      & 0       & \lambda \\\end{pmatrix}
=
\lambda \begin{pmatrix}
1 & \lambda^{-1}       & 0       & 0      & \cdots  & 0 \\
0       & 1 & \lambda^{-1}       & 0      & \cdots  & 0 \\
0       & 0       & 1 & \lambda^{-1}      & \cdots  & 0 \\
\vdots  & \vdots  & \vdots  & \ddots & \ddots  & \vdots \\
0       & 0       & 0       & 0      & 1 & \lambda^{-1}       \\
0       & 0       & 0       & 0      & 0       & 1 \\\end{pmatrix}=\lambda(I+K)

where K is a matrix with zeros on and under the main diagonal. (The number λ is nonzero by the assumption that the matrix whose logarithm one attempts to take is invertible.)

Then, by the formula

 \ln (1+x)=x-\frac{x^2}{2}+\frac{x^3}{3}-\frac{x^4}{4}+\cdots

one gets

\ln B=\ln \big(\lambda(I+K)\big)=\ln (\lambda I) +\ln (I+K)= (\ln \lambda) I + K-\frac{K^2}{2}+\frac{K^3}{3}-\frac{K^4}{4}+\cdots

This series in general does not converge for every matrix K, as it would not for any real number with absolute value greater than unity, however, this particular K is a nilpotent matrix, so the series actually has a finite number of terms (Km is zero if m is the dimension of K).

Using this approach one finds

\ln \begin{bmatrix}1 & 1\\ 0 & 1\end{bmatrix}
=\begin{bmatrix}0 & 1\\ 0 & 0\end{bmatrix}.

A functional analysis perspective

A square matrix represents a linear operator on the Euclidean space Rn where n is the dimension of the matrix. Since such a space is finite-dimensional, this operator is actually bounded.

Using the tools of holomorphic functional calculus, given a holomorphic function f(z) defined on an open set in the complex plane and a bounded linear operator T, one can calculate f(T) as long as f(z) is defined on the spectrum of T.

The function f(z)=ln z can be defined on any simply connected open set in the complex plane not containing the origin, and it is holomorphic on such a domain. This implies that one can define ln T as long as the spectrum of T does not contain the origin and there is a path going from the origin to infinity not crossing the spectrum of T (as such, if the spectrum of T is a circle with the origin inside of it, it is impossible to define ln T).

Back to the particular case of a Euclidean space, the spectrum of a linear operator on this space is the set of eigenvalues of its matrix, and so is a finite set. As long as the origin is not in the spectrum (the matrix is invertible), one obviously satisfies the path condition from the previous paragraph, and as such, the theory implies that ln T is well-defined. The non-uniqueness of the matrix logarithm then follows from the fact that one can choose more than one branch of the logarithm which is defined on the set of eigenvalues of a matrix.

A Lie group theory perspective

In the theory of Lie groups, there is an exponential map from a Lie algebra g to the corresponding Lie group G

 \exp : g \rightarrow G.

For matrix Lie groups, the elements of g and G are square matrices and the exponential map is given by the matrix exponential. The inverse map log  = exp  − 1 is multivalued and coincides with the matrix logarithm discussed here. The logarithm maps from the Lie group G into the Lie algebra g. Note that the exponential map is a local diffeomorphism between a neighborhood U of the zero matrix  \underline{0} \in g and a neighborhood V of the identity matrix \underline{1}\in G. Thus the (matrix) logarithm is well-defined as a map

 \log: V\subset G\rightarrow U\subset g. \,

Constraints in 2 × 2 case

If a 2 x 2 real matrix has a negative determinant, it has no real logarithm. Note first that any 2 × 2 real matrix can be considered one of the three types of complex number z = x + y ε where \epsilon^2 = -1, 0 , or + 1. This z is a point on a complex subplane of the ring of matrices. The case where the determinant is negative only arises in a plane with ε2 = + 1, that is a split-complex number plane. Only one quarter of this plane is the image of the exponential map, so logarithm is only defined on that quarter (quadrant). The other three quadrants are images of this one under the Klein four-group generated by ε and −1.

For example, let a = ln 2, then cosh a = 5/4 and sinh a = 3/4. For matrices this means \exp \begin{pmatrix}0 & a \\ a & 0 \end{pmatrix} = 
\begin{pmatrix}\cosh a & \sinh a \\ \sinh a & \cosh a  \end{pmatrix} =
\begin{pmatrix}1.25 & .75\\ .75 & 1.25 \end{pmatrix}. So this last matrix has logarithm \begin{pmatrix}0 & \ln 2 \\ \ln 2 & 0 \end{pmatrix}.

These matrices do not have a logarithm: \begin{pmatrix}3/4 & 5/4 \\ 5/4 & 3/4 \end{pmatrix},\ 
\begin{pmatrix}-3/4 & -5/4 \\ -5/4 & -3/4\end{pmatrix}, \ 
\begin{pmatrix}-5/4 & -3/4\\ -3/4 & -5/4 \end{pmatrix}. They represent three other conjugates by the four-group of the matrix above that does have a logarithm. A non-singular 2 x 2 matrix does not necessarily have a logarithm, but it is conjugate by the four-group to a matrix that does have a logarithm.

For a richer affirmative example, start with a pythagorean triple (p,q,r) and let a = ln(p + r) − ln q. Then

e^a = \frac {p + r} {q} = \cosh a + \sinh a.

Now \exp \begin{pmatrix}0 & a \\ a & 0 \end{pmatrix} =
\begin{pmatrix}r/q & p/q \\ p/q & r/q \end{pmatrix}. Thus \tfrac{1}{q}\begin{pmatrix}r & p \\ p & r \end{pmatrix} has logarithm matrix \begin{pmatrix}0 & a \\ a & 0 \end{pmatrix} where a = ln(p + q) − ln q.

See also

Notes

  1. ^ Higham (2008), Theorem 1.27
  2. ^ Higham (2008), Theorem 1.31
  3. ^ Culver (1966)

References


Wikimedia Foundation. 2010.

Игры ⚽ Нужна курсовая?

Look at other dictionaries:

  • Logarithm — The graph of the logarithm to base 2 crosses the x axis (horizontal axis) at 1 and passes through the points with coordinates (2, 1), (4, 2), and (8, 3) …   Wikipedia

  • Matrix exponential — In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. Abstractly, the matrix exponential gives the connection between a matrix Lie algebra and the corresponding Lie group.… …   Wikipedia

  • Matrix function — In mathematics, a matrix function is a function which maps a matrix to another matrix. Contents 1 Extending scalar functions to matrix functions 1.1 Power series 1.2 Jordan decomposition …   Wikipedia

  • Square root of a matrix — In mathematics, the square root of a matrix extends the notion of square root from numbers to matrices. A matrix B is said to be a square root of A if the matrix product B · B is equal to A.[1] Contents 1 Properties 2 Computation methods …   Wikipedia

  • Natural logarithm — Base e redirects here. For the numbering system which uses e as its base, see Non integer representation#Base e. Graph of the natural logarithm function. The function slowly grows to positive infinity as x increases and rapidly goes to negative… …   Wikipedia

  • Jordan matrix — In the mathematical discipline of matrix theory, a Jordan block over a ring R (whose identities are the zero 0 and one 1) is a matrix which is composed of 0 elements everywhere except for the diagonal, which is filled with a fixed element… …   Wikipedia

  • Super-logarithm — In mathematics, the super logarithm is one of the two inverse functions of tetration. Just as exponentiation has two inverse functions: roots and logarithms, likewise tetration has two inverse functions: super roots and super logarithms. There… …   Wikipedia

  • Substitution matrix — In evolutionary biology, a substitution matrix describes the rate at which one character in a sequence changes to other character states over time. Substitution matrices are usually seen in the context of amino acid or DNA sequence alignments,… …   Wikipedia

  • Density matrix — Mixed state redirects here. For the psychiatric condition, see Mixed state (psychiatry). In quantum mechanics, a density matrix is a self adjoint (or Hermitian) positive semidefinite matrix (possibly infinite dimensional) of trace one, that… …   Wikipedia

  • Hilbert matrix — In linear algebra, a Hilbert matrix is a matrix with the unit fraction elements: H {ij} = frac{1}{i+j 1}. For example, this is the 5 times; 5 Hilbert matrix::H = egin{bmatrix} 1 frac{1}{2} frac{1}{3} frac{1}{4} frac{1}{5} [4pt] frac{1}{2}… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”