Positive-definite matrix

Positive-definite matrix

In linear algebra, a positive-definite matrix is a matrix that in many ways is analogous to a positive real number. The notion is closely related to a positive-definite symmetric bilinear form (or a sesquilinear form in the complex case).

The proper definition of positive-definite is unambiguous for Hermitian matrices, but there is no agreement in the literature on how this should be extended for non-Hermitian matrices, if at all. (See the section Non-Hermitian matrices below.)



An n × n real matrix M is positive definite if zTMz > 0 for all non-zero vectors z with real entries (z \in \mathbb{R}^n), where zT denotes the transpose of z.

An n × n complex matrix M is positive definite if ℜ(z*Mz) > 0 for all non-zero complex vectors z, where z* denotes the conjugate transpose of z and ℜ(c) is the real part of a complex number c.

An n × n complex Hermitian matrix M is positive definite if z*Mz > 0 for all non-zero complex vectors z. The quantity z*Mz is always real because M is a Hermitian matrix.


  • The matrix  M_0 =  \begin{bmatrix} 1 & 0 \\ 0 & 1\end{bmatrix} is positive definite. For a vector with entries \textbf{z}= \begin{bmatrix} z_0 \\ z_1\end{bmatrix} the quadratic form is  \begin{bmatrix} z_0 & z_1\end{bmatrix} \begin{bmatrix} 1 & 0 \\ 0 & 1 \end{bmatrix} \begin{bmatrix} z_0 \\ z_1\end{bmatrix}=\begin{bmatrix} z_0\cdot 1+z_1\cdot 0 & z_0\cdot 0+z_1\cdot 1\end{bmatrix} \begin{bmatrix} z_0 \\ z_1\end{bmatrix}=z_0^2+z_1^2;

when the entries z0, z1 are real and at least one of them nonzero, this is positive.

  • An example of a matrix that is not positive, but is positive-definite, is given by
 M_1 = \begin{bmatrix} 2&-1&0\\-1&2&-1\\0&-1&2 \end{bmatrix}.

It is positive definite since for any non-zero vector  x = \begin{bmatrix} x_1\\x_2\\x_3 \end{bmatrix} , we have

x^{\mathrm T}M_1 x 
&= \begin{bmatrix} x_1&x_2&x_3 \end{bmatrix} \begin{bmatrix} 2&-1&0\\-1&2&-1\\0&-1&2 \end{bmatrix} \begin{bmatrix} x_1\\x_2\\x_3 \end{bmatrix} \\
&= \begin{bmatrix} (2x_1-x_2)&(-x_1+2x_2-x_3)&(-x_2+2x_3) \end{bmatrix} \begin{bmatrix} x_1\\x_2\\x_3 \end{bmatrix} \\
&= 2{x_1}^2 - 2x_1x_2 + 2{x_2}^2 - 2x_2x_3 + 2{x_3}^2 \\
&= {x_1}^2+(x_1 - x_2)^{2} + (x_2 - x_3)^{2}+{x_3}^2

which is a sum of squares and therefore nonnegative; in fact, each squared summand can be zero only when x1 = x2 = x3 = 0, so M1 is indeed positive-definite.

  • Conversely, the strictly positive matrix M_2 =  \begin{bmatrix} 1 & 2 \\ 2 & 1\end{bmatrix} is not positive definite, showing (together with the previous example) that these two properties are independent. Evaluated at \textbf{z}= \begin{bmatrix} 1\\ -1\end{bmatrix}, the quadratic form is  \begin{bmatrix} 1 & -1\end{bmatrix} \begin{bmatrix} 1 & 2 \\ 2 & 1 \end{bmatrix} \begin{bmatrix} 1 \\ -1\end{bmatrix}=-2 < 0.
  • For a huge class of examples, consider that in statistics positive definite matrices appear as covariance matrices. In fact, every positive definite matrix is the covariance matrix of some multivariate probability distribution.


Let M be an n × n Hermitian matrix. The following properties are equivalent to M being positive definite:

1. All eigenvalues λi of M are positive. Recall that any Hermitian M has an eigendecomposition M = P−1DP where P is a unitary matrix whose rows are orthonormal eigenvectors of M, forming a basis, and D is a diagonal matrix. Therefore M may be regarded as a real diagonal matrix D that has been re-expressed in some new coordinate system. This characterization means that M is positive definite if and only if the diagonal elements of D (the eigenvalues) are all positive. In other words, in the basis consisting of the eigenvectors of M, the action of M is component-wise multiplication with a (fixed) element in Cn with positive entries[clarification needed].
2. The sesquilinear form
\langle \mathbf{x},\mathbf{y}\rangle := \mathbf{y}^*\mathbf{M}\mathbf{x}

defines an inner product on Cn. (In fact, every inner product on Cn arises in this fashion from a Hermitian positive definite matrix.) In particular, positive definiteness for a Hermitian M is equivalent to the fact that \langle \mathbf{x},\mathbf{x}\rangle > 0 for all nonzero x.

3. M is the Gram matrix of some collection of linearly independent vectors
\textbf{x}_1,\ldots,\textbf{x}_n \in \mathbb{C}^k

for some k. That is, M satisfies:

M_{ij} = \langle \textbf{x}_i, \textbf{x}_j\rangle = \textbf{x}_i^{*} \textbf{x}_j.

The vectors xi may optionally be restricted to fall in Cn. In other words, M is of the form A * A where A is not necessarily square but must be injective in general.

4. All the following matrices have a positive determinant (Sylvester's criterion):
  • the upper left 1-by-1 corner of M
  • the upper left 2-by-2 corner of M
  • the upper left 3-by-3 corner of M
  • ...
  • M itself

In other words, all of the leading principal minors are positive. For positive semidefinite matrices, all principal minors have to be non-negative. The leading principal minors alone do not imply positive semidefiniteness, as can be seen from the example

 \begin{bmatrix} 0 & 0 \\ 0 & -1 \end{bmatrix}.
5. There exists a unique lower triangular matrix L, with strictly positive diagonal elements, that allows the factorization of M into
M = LL * .

where L * is the conjugate transpose of L. This factorization is called Cholesky decomposition.

6. The quadratic function associated with M
 f(\mathbf{x}) = \frac12 \mathbf{x}^\mathrm{T} M \mathbf{x} - \mathbf{x}^\mathrm{T} \mathbf{b}

is, regardless of b, a strictly convex function. f(\mathbf{x}) has a unique global minimum, and this is why positive definite matrices are so common in optimization problems.

For real symmetric matrices, these properties can be simplified by replacing \mathbb{C}^n with \mathbb{R}^n, and "conjugate transpose" with "transpose."

Quadratic forms

Echoing condition 2 above, one can also formulate positive-definiteness in terms of quadratic forms. Let K be the field R or C, and V be a vector space over K. A Hermitian form

B : V \times V \rightarrow K

is a bilinear map such that B(x, y) is always the complex conjugate of B(y, x). Such a function B is called positive definite if B(x, x) > 0 for every nonzero x in V.

Simultaneous diagonalization

Two symmetric, positive-definite matrices can be simultaneously diagonalized, although not necessarily via a similarity transformation. This result does not extend to the case of three or more matrices. In this section we write for the real case, extension to the complex case is immediate.

Let M and N be two positive-definite matrices. Write the generalized eigenvalue equation as (M − λN)x = 0 where xTNx = 1. Now we can decompose the inverse of N as QTQ (so N must be positive definite, as the proof shows, in fact it is enough that M is symmetric). Now multiply various places with Q to get Q(M − λN)QTQ − Tx = 0 which we can rewrite as (QMQT)y = λy where yTy = 1. Manipulation now yields MX = NXΛ where X is a matrix having as columns the generalized eigenvectors and Λ is a diagonal matrix with the generalized eigenvalues. Now premultiplication with XT gives the final result:

XTMX = Λ and XTNX = I, but note that this is no longer an orthogonal diagonalization.

Note that this result does not contradict what is said on simultaneous diagonalization in the article Diagonalizable matrix#Simultaneous diagonalisation, which refers to simultaneous diagonalization by a similarity transformation. Our result here is more akin to a simultaneous diagonalization of two quadratic forms, and is useful for optimization of one form under conditions on the other. For this result see Horn&Johnson, 1985, page 218 and following.

Negative-definite, semidefinite and indefinite matrices

The n × n Hermitian matrix M is said to be negative-definite if

x^{*} M x < 0\,

for all non-zero x \in \mathbb{C}^n (or, all non-zero x \in \mathbb{R}^n for the real matrix).

It is called positive-semidefinite (or sometimes nonnegative-definite) if

x^{*} M x \geq 0

for all x \in \mathbb{C}^n (or, all x \in \mathbb{R}^n for the real matrix), where x * is the conjugate transpose of x.

It is called negative-semidefinite if

x^{*} M x \leq 0

for all x \in \mathbb{C}^n (or, all x \in \mathbb{R}^n for the real matrix).

A matrix M is positive-semidefinite if and only if it arises as the Gram matrix of some set of vectors. In contrast to the positive-definite case, these vectors need not be linearly independent.

For any matrix A, the matrix A*A is positive semidefinite, and rank(A) = rank(A*A). Conversely, any Hermitian positive semidefinite matrix M can be written as M = A*A; this is the Cholesky decomposition.

A Hermitian matrix which is neither positive- nor negative-semidefinite is called indefinite.

A matrix is negative definite if all kth order leading principal minors are negative if k is odd and positive if k is even.

A Hermitian matrix is negative-definite, negative-semidefinite, or positive-semidefinite if and only if all of its eigenvalues are negative, non-positive, or non-negative, respectively.

Further properties

If M is an Hermitian positive-semidefinite matrix, one sometimes writes M ≥ 0 and if M is positive-definite one writes M > 0.[1] The notion comes from functional analysis where positive definite matrices define positive operators.

For arbitrary square matrices M,N we write M ≥ N if M − N ≥ 0; i.e., M − N is positive semi-definite. This defines a partial ordering on the set of all square matrices. One can similarly define a strict partial ordering M > N.

  1. Every positive definite matrix is invertible and its inverse is also positive definite.[2] If M ≥ N > 0 then N−1 ≥ M−1 > 0.[3]
  2. If M is positive definite and r > 0 is a real number, then rM is positive definite.[4] If M and N are positive definite, then the sum M + N[4] and the products MNM and NMN are also positive definite. If MN = NM, then MN is also positive definite.
  3. If M = (mij) ≥ 0 then the diagonal entries mii are real and non-negative. As a consequence the trace, tr(M) ≥ 0. Furthermore[5]
      |m_{ij}| \leq \sqrt{m_{ii} m_{jj}} \leq \frac{1}{2}(m_{ii}+m_{jj})
    and thus
     \max |m_{ij}| \leq \max|m_{ii}|
  4. If M is a symmetric matrix of the form mij = m(ij), and the strict inequality holds
\sum_{j\neq 0}|m(j)|< m(0)
    then M is strictly positive definite.
  5. A matrix M is positive semi-definite if and only if there is a positive semi-definite matrix B with B2 = M. This matrix B is unique,[6] is called the square root of M, and is denoted with B = M1/2 (the square root B is not to be confused with the matrix L in the Cholesky factorization M = LL*, which is also sometimes called the square root of M). If M > N > 0 then M1/2 > N1/2 > 0.
  6. If M,N > 0 then M ⊗ N > 0, where ⊗ denotes Kronecker product.
  7. For matrices M = (mij) and N = (nij), write M ° N for the entry-wise product of M and N; i.e., the matrix whose i,j entry is mijnij. Then M ° N is the Hadamard product of M and N. The Hadamard product of two positive-definite matrices is again positive-definite and the Hadamard product of two positive-semidefinite matrices is again positive-semidefinite (this result is often called the Schur product theorem).[7] Furthermore, if M and N are positive-semidefinite, then the following inequality, due to Oppenheim, holds:[8]
    \det(M\cdot N) \geq (\det N) \prod_{i} m_{ii}.
  8. Let M > 0 and N Hermitian. If MN + NM ≥ 0 (resp., MN + NM > 0) then N ≥ 0 (resp., N > 0).[citation needed]
  9. If M,N ≥ 0 are real positive semidefinite symmetric matrices, then \scriptstyle\operatorname{tr}(MN) \,\geq\, 0 (Lancaster-Tismenetsky, The Theory of Matrices, p. 218).
  10. If M > 0 is real, then there is a δ > 0 such that M > δI, where I is the identity matrix.

Block Matrices

A positive 2n × 2n matrix may also be defined by blocks:

 M =  \begin{bmatrix} A & B \\ C & D \end{bmatrix}

Where each block is n × n. By applying the positivity condition, it immediately follows that A and D are hermitian, and C = B*.

We have that z*Mz ≥ 0 for all complex z, and in particular for z = ( v, 0)T. Then

 \begin{bmatrix} v^* & 0  \end{bmatrix} \begin{bmatrix} A & B \\ B^* & D \end{bmatrix} \begin{bmatrix} v \\ 0 \end{bmatrix} = v^* A v \ge 0

A similar argument can be applied to D, and thus we conclude that both A and D must be positive definite matrices, as well.

Non-Hermitian matrices

A real matrix M may have the property that xTMx > 0 for all nonzero real vectors x without being symmetric. The matrix

 \begin{bmatrix} 1 & 1 \\ -1 & 1 \end{bmatrix}

satisfies this property, because for all real vectors x = (x1,x2)T such that x \ne 0,

 \begin{bmatrix} x_1 & x_2 \end{bmatrix} \begin{bmatrix} 1 & 1 \\ -1 & 1 \end{bmatrix} \begin{bmatrix} x_1 \\ x_2 \end{bmatrix} = x_1^2 + x_2^2 > 0 .

In general, we have xTMx > 0 for all real nonzero vectors x if and only if the symmetric part, (M + MT) / 2, is positive definite.

The situation for complex matrices may be different, depending on how one generalizes the inequality z*Mz > 0 when considering M which aren't necessarily Hermitian. If z*Mz is real for all complex vectors z, then the matrix M must be Hermitian. To see this, we define the Hermitian matrices A=(M+M*)/2 and B=(M-M*)/(2i) so that M=A+iB. Then, z*Mz=z*Az+iz*Bz is real. By the Hermiticity of A and B, z*Az and z*Bz are individually real so z*Bz must be zero for all z. So then B is the zero matrix and M=A, which is Hermitian.

So, if we require that z*Mz be real and positive, then M is automatically Hermitian. On the other hand, we have that Re(z*Mz) > 0 for all complex nonzero vectors z if and only if the Hermitian part, (M + M*) / 2, is positive definite.

In summary, the distinguishing feature between the real and complex case is that, a bounded positive operator on a complex Hilbert space is necessarily Hermitian, or self adjoint. The general claim can be argued using the polarization identity. That is no longer true in the real case.

There is no agreement in the literature on the proper definition of positive-definite for non-Hermitian matrices.

See also


  1. ^ This may be confusing, as sometimes nonnegative matrices are also denoted in this way. A common alternative notation is M \succeq 0 and  M \succ 0 for positive semidefinite and positive definite matrices, respectively.
  2. ^ Horn & Johnson (1985), p. 397
  3. ^ Horn & Johnson (1985), Corollary 7.7.4(a)
  4. ^ a b Horn & Johnson (1985), Observation 7.1.3
  5. ^ Horn & Johnson (1985), p. 398
  6. ^ Horn & Johnson (1985), Theorem 7.2.6 with k = 2
  7. ^ Horn & Johnson (1985), Theorem 7.5.3
  8. ^ Horn & Johnson (1985), Theorem 7.8.6


External links

Wikimedia Foundation. 2010.

Look at other dictionaries:

  • Matrix - получить на Академике рабочий купон на скидку Строительный Двор или выгодно matrix купить с бесплатной доставкой на распродаже в Строительный Двор

  • Positive definite — In mathematics, positive definite may refer to: * positive definite matrix * positive definite function ** positive definite function on a group * positive definite bilinear form …   Wikipedia

  • positive definite — positive definiteness. Math. 1. (of a quadratic form) positive for all real values of the variables, where the values are not all zero. 2. (of a matrix) displaying the coefficients of a positive definite quadratic form. [1905 10] * * * …   Universalium

  • Positive-definite function — In mathematics, the term positive definite function may refer to a couple of different concepts. Contents 1 In dynamical systems 2 In analysis 2.1 Bochner s theorem 2.1.1 Applications …   Wikipedia

  • Positive definite kernel — In operator theory, a positive definite kernel is a generalization of a positive matrix. Definition Let :{ H n } {n in {mathbb Z be a sequence of (complex) Hilbert spaces and :mathcal{L}(H i, H j)be the bounded operators from Hi to Hj . A map A… …   Wikipedia

  • Positive definite function on a group — In operator theory, a positive definite function on a group relates the notions of positivity, in the context of Hilbert spaces, and algebraic groups. It can be viewed as a particular type of positive definite kernel where the underlying set has… …   Wikipedia

  • positive definite — adjective 1. : having a positive value of all values of the constituent variables positive definite quadratic forms 2. of a matrix : having the characteristic roots real and positive * * * positive definiteness. Math. 1. (of a quadratic form)… …   Useful english dictionary

  • positive definite — adjective Date: 1907 1. having a positive value for all values of the constituent variables < positive definite quadratic forms > 2. of a matrix having the characteristic roots real and positive …   New Collegiate Dictionary

  • Positiv definite Matrix — Definitheit ist ein Begriff aus dem mathematischen Teilgebiet der linearen Algebra. Er beschreibt, welche Vorzeichen reelle quadratische Formen annehmen können, die durch Matrizen oder allgemeiner durch Bilinearformen erzeugt werden.… …   Deutsch Wikipedia

  • Positive definiteness — is a property of the following mathematical objects:* Positive definite bilinear form * Positive definite matrix * Positive definite function …   Wikipedia

  • Matrix decomposition — In the mathematical discipline of linear algebra, a matrix decomposition is a factorization of a matrix into some canonical form. There are many different matrix decompositions; each finds use among a particular class of problems. Contents 1… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”