Diagonally dominant matrix

Diagonally dominant matrix

In mathematics, a matrix is said to be diagonally dominant if for every row of the matrix, the magnitude of the diagonal entry in a row is larger than or equal to the sum of the magnitudes of all the other (non-diagonal) entries in that row. More precisely, the matrix A is diagonally dominant if

|a_{ii}| \geq \sum_{j\neq i} |a_{ij}| \quad\text{for all } i, \,

where aij denotes the entry in the ith row and jth column.

Note that this definition uses a weak inequality, and is therefore sometimes called weak diagonal dominance. If a strict inequality (>) is used, this is called strict diagonal dominance. The unqualified term diagonal dominance can mean both strict and weak diagonal dominance, depending on the context.[1]

Contents

Variations

The definition in the first paragraph sums entries across rows. It is therefore sometimes called row diagonal dominance. If one changes the definition to sum down columns, this is called column diagonal dominance.

If an irreducible matrix is weakly diagonally dominant, but in at least one row (or column) is strictly diagonally dominant, then the matrix is irreducibly diagonally dominant.

Examples

The matrix

\mathbf A = \begin{bmatrix}
3 & -2 & 1\\
1 & -3 & 2\\
-1 & 2 & 4\end{bmatrix}

gives

|a_{11}| \ge |a_{12}| + |a_{13}|   since   |3| \ge |-2| + |1|
|a_{22}| \ge |a_{21}| + |a_{23}|   since   |-3| \ge |1| + |2|
|a_{33}| \ge |a_{31}| + |a_{32}|   since   |4| \ge |-1| + |2|.

Because the magnitude of each diagonal element is greater than or equal to the sum of the magnitude of other elements in the row, A is diagonally dominant.

The matrix

\mathbf B = \begin{bmatrix}
-2 & 2 & 1\\
1 & 3 & 2\\
1 & -2 & 0\end{bmatrix}

But here,

| b11 | < | b12 | + | b13 |   since   | − 2 | < | 2 | + | 1 |
|b_{22}| \ge |b_{21}| + |b_{23}|   since   |3| \ge |1| + |2|
| b33 | < | b31 | + | b32 |   since   | 0 | < | 1 | + | − 2 | .

Because | b11 | and | b33 | are less than the sum of the magnitude of other elements in their respective row, B is not diagonally dominant.

The matrix

\mathbf C = \begin{bmatrix}
-4 & 2 & 1\\
1 & 6 & 2\\
1 & -2 & 5\end{bmatrix}

gives

|c_{11}| \ge |c_{12}| + |c_{13}|   since   | − 4 | > | 2 | + | 1 |
|c_{22}| \ge |c_{21}| + |c_{23}|   since   | 6 | > | 1 | + | 2 |
|c_{33}| \ge |c_{31}| + |c_{32}|   since   | 5 | > | 1 | + | − 2 | .

Because the magnitude of each diagonal element is greater than the sum of the magnitude of the other elements in the row, C is strictly diagonally dominant.

Applications and properties

By the Gershgorin circle theorem, a strictly (or irreducibly) diagonally dominant matrix is non-singular. This result is known as the Levy–Desplanques theorem.[2]

A Hermitian diagonally dominant matrix with real non-negative diagonal entries is positive semi-definite. If the symmetry requirement is eliminated, such a matrix is not necessarily positive semi-definite; however, the real parts of its eigenvalues are non-negative.

No (partial) pivoting is necessary for a strictly column diagonally dominant matrix when performing Gaussian elimination (LU factorization).

The Jacobi and Gauss–Seidel methods for solving a linear system converge if the matrix is strictly (or irreducibly) diagonally dominant. In October 2010, researchers at Carnegie Mellon University announced that they have discovered a new algorithm for solving symmetric diagonally dominant linear systems.[3]

Many matrices that arise in finite element methods are diagonally dominant.

A slight variation on the idea of diagonal dominance is used to prove that the pairing on diagrams without loops in the Temperley–Lieb algebra is nondegenerate.[4] For a matrix with polynomial entries, one sensible definition of diagonal dominance is if the highest power of q appearing in each row appears only on the diagonal. (The evaluations of such a matrix at large values of q are diagonally dominant in the above sense.)

Links

Notes

  1. ^ For instance, Horn and Johnson (1985, p. 349) use it to mean weak diagonal dominance.
  2. ^ Horn and Johnson, Thm 6.1.10. This result has been independently rediscovered dozens of times. A few notable ones are Lévy (1881), Desplanques (1886), Minkowski (1900), Hadamard (1903), Schur, Markov (1908), Rohrbach (1931), Gershgorin (1931), Artin (1932), Ostrowski (1937), and Furtwängler (1936). For a history of this "recurring theorem" see: Taussky, Olga (1949). "A recurring theorem on determinants". American Mathematical Monthly (The American Mathematical Monthly, Vol. 56, No. 10) 56 (10): 672–676. doi:10.2307/2305561. JSTOR 2305561.  Another useful history is in: Schneider, Hans (1977). "Olga Taussky-Todd's influence on matrix theory and matrix theorists". Linear and Multilinear Algebra 5 (3): 197–224. doi:10.1080/03081087708817197. 
  3. ^ http://www.pcpro.co.uk/news/362170/algorithm-sees-massive-leap-in-complex-number-crunching
  4. ^ K.H. Ko and L. Smolinski (1991). "A combinatorial matrix in 3-manifold theory". Pacific. J. Math. 149: 319–336. 

References

  • Gene H. Golub & Charles F. Van Loan. Matrix Computations, 1996. ISBN 0-8018-5414-8
  • Roger A. Horn & Charles R. Johnson. Matrix Analysis, Cambridge University Press, 1985. ISBN 0-521-38632-2 (paperback).

Wikimedia Foundation. 2010.

Игры ⚽ Поможем написать курсовую

Look at other dictionaries:

  • Diagonal matrix — In linear algebra, a diagonal matrix is a matrix (usually a square matrix) in which the entries outside the main diagonal (↘) are all zero. The diagonal entries themselves may or may not be zero. Thus, the matrix D = (di,j) with n columns and n… …   Wikipedia

  • List of matrices — This page lists some important classes of matrices used in mathematics, science and engineering: Matrices in mathematics*(0,1) matrix a matrix with all elements either 0 or 1. Also called a binary matrix . *Adjugate matrix * Alternant matrix a… …   Wikipedia

  • List of numerical analysis topics — This is a list of numerical analysis topics, by Wikipedia page. Contents 1 General 2 Error 3 Elementary and special functions 4 Numerical linear algebra …   Wikipedia

  • Список матриц — Структура матрицы Здесь собраны наиболее важные классы матриц, используемые в математике, науке (в целом) и прикладной науке (в частности). Под матрицей понимается прямоугольный массив чисел …   Википедия

  • List of mathematics articles (D) — NOTOC D D distribution D module D D Agostino s K squared test D Alembert Euler condition D Alembert operator D Alembert s formula D Alembert s paradox D Alembert s principle Dagger category Dagger compact category Dagger symmetric monoidal… …   Wikipedia

  • Interval arithmetic — Interval arithmetic, also called interval mathematics , interval analysis , and interval computation , is a method in mathematics. It has been developed by mathematicians since the 1950s and 1960s as an approach to putting bounds on rounding… …   Wikipedia

  • Constraint algorithm — In mechanics, a constraint algorithm is a method for satisfying constraints for bodies that obey Newton s equations of motion. There are three basic approaches to satisfying such constraints: choosing novel unconstrained coordinates ( internal… …   Wikipedia

  • Gauss–Seidel method — The Gauss–Seidel method is a technique used to solve a linear system of equations. The method is named after the German mathematicians Carl Friedrich Gauss and Philipp Ludwig von Seidel. The method is an improved version of the Jacobi method. It… …   Wikipedia

  • Gaussian elimination — In linear algebra, Gaussian elimination is an algorithm for solving systems of linear equations. It can also be used to find the rank of a matrix, to calculate the determinant of a matrix, and to calculate the inverse of an invertible square… …   Wikipedia

  • Jacobi method — The Jacobi method is an algorithm in linear algebra for determining the solutions of a system of linear equations with largest absolute values in each row and column dominated by the diagonal element. Each diagonal element is solved for, and an… …   Wikipedia

Share the article and excerpts

Direct link
Do a right-click on the link above
and select “Copy Link”