Diagonal Matrix
Get Diagonal Matrix essential facts below. View Videos or join the Diagonal Matrix discussion. Add Diagonal Matrix to your PopFlock.com topic list for future reference or share this resource on social media.
Diagonal Matrix

In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. An example of a 2-by-2 diagonal matrix is ${\displaystyle \left[{\begin{smallmatrix}3&0\\0&2\end{smallmatrix}}\right]}$, while an example of a 3-by-3 diagonal matrix is${\displaystyle \left[{\begin{smallmatrix}6&0&0\\0&7&0\\0&0&4\end{smallmatrix}}\right]}$. An identity matrix of any size, or any multiple of it (a scalar matrix), is a diagonal matrix.

A diagonal matrix is sometimes called a scaling matrix, since matrix multiplication with it results in changing scale (size). Its determinant is the product of its diagonal values.

## Definition

As stated above, a diagonal matrix is a matrix in which all off-diagonal entries are zero. That is, the matrix with n columns and n rows is diagonal if

${\displaystyle \forall i,j\in \{1,2,\ldots ,n\},i\neq j\implies d_{i,j}=0}$.

However, the main diagonal entries are unrestricted.

The term diagonal matrix may sometimes refer to a rectangular diagonal matrix, which is an m-by-n matrix with all the entries not of the form di,i being zero. For example:

${\displaystyle {\begin{bmatrix}1&0&0\\0&4&0\\0&0&-3\\0&0&0\\\end{bmatrix}}}$ or ${\displaystyle {\begin{bmatrix}1&0&0&0&0\\0&4&0&0&0\\0&0&-3&0&0\end{bmatrix}}}$

More often, however, diagonal matrix refers to square matrices, which can be specified explicitly as a square diagonal matrix. A square diagonal matrix is a symmetric matrix, so this can also be called a symmetric diagonal matrix.

The following matrix is square diagonal matrix:

${\displaystyle {\begin{bmatrix}1&0&0\\0&4&0\\0&0&-2\end{bmatrix}}}$

If the entries are real numbers or complex numbers, then it is a normal matrix as well.

In the remainder of this article we will consider only square diagonal matrices, and refer to them simply as "diagonal matrices".

## Scalar matrix

A diagonal matrix with all its main diagonal entries equal is a scalar matrix, that is, a scalar multiple λI of the identity matrix I. Its effect on a vector is scalar multiplication by λ. For example, a 3×3 scalar matrix has the form:

${\displaystyle {\begin{bmatrix}\lambda &0&0\\0&\lambda &0\\0&0&\lambda \end{bmatrix}}\equiv \lambda {\boldsymbol {I}}_{3}}$

The scalar matrices are the center of the algebra of matrices: that is, they are precisely the matrices that commute with all other square matrices of the same size.[a] By contrast, over a field (like the real numbers), a diagonal matrix with all diagonal elements distinct only commutes with diagonal matrices (its centralizer is the set of diagonal matrices). That is because if a diagonal matrix ${\displaystyle D=\mathrm {diag} (a_{1},\dots ,a_{n})}$ has ${\displaystyle a_{i}\neq a_{j},}$ then given a matrix ${\displaystyle M}$ with ${\displaystyle m_{ij}\neq 0,}$ the ${\displaystyle (i,j)}$ term of the products are: ${\displaystyle (DM)_{ij}=a_{j}m_{ij}}$ and ${\displaystyle (MD)_{ij}=m_{ij}a_{i},}$ and ${\displaystyle a_{j}m_{ij}\neq m_{ij}a_{i}}$ (since one can divide by ${\displaystyle m_{ij}}$), so they do not commute unless the off-diagonal terms are zero.[b] Diagonal matrices where the diagonal entries are not all equal or all distinct have centralizers intermediate between the whole space and only diagonal matrices.[1]

For an abstract vector space V (rather than the concrete vector space ${\displaystyle K^{n}}$), or more generally a module M over a ring R, with the endomorphism algebra End(M) (algebra of linear operators on M) replacing the algebra of matrices, the analog of scalar matrices are scalar transformations. Formally, scalar multiplication is a linear map, inducing a map ${\displaystyle R\to \operatorname {End} (M),}$ (send a scalar λ to the corresponding scalar transformation, multiplication by λ) exhibiting End(M) as a R-algebra. For vector spaces, or more generally free modules ${\displaystyle M\cong R^{n}}$, for which the endomorphism algebra is isomorphic to a matrix algebra, the scalar transforms are exactly the center of the endomorphism algebra, and similarly invertible transforms are the center of the general linear group GL(V), where they are denoted by Z(V), follow the usual notation for the center.

## Vector operations

Multiplying a vector by a diagonal matrix multiplies each of the terms by the corresponding diagonal entry. Given a diagonal matrix ${\displaystyle D=\mathrm {diag} (a_{1},\dots ,a_{n})}$ and a vector ${\displaystyle v=\left[{\begin{smallmatrix}x_{1}\\\vdots \\x_{n}\end{smallmatrix}}\right]}$, the product is:

${\displaystyle Dv=\mathrm {diag} (a_{1},\dots ,a_{n}){\begin{bmatrix}x_{1}\\\vdots \\x_{n}\end{bmatrix}}={\begin{bmatrix}a_{1}\\&\ddots \\&&a_{n}\end{bmatrix}}{\begin{bmatrix}x_{1}\\\vdots \\x_{n}\end{bmatrix}}={\begin{bmatrix}a_{1}x_{1}\\\vdots \\a_{n}x_{n}\end{bmatrix}}.}$

This can be expressed more compactly by using a vector instead of a diagonal matrix, ${\displaystyle d=\left[{\begin{smallmatrix}a_{1}\\\vdots \\a_{n}\end{smallmatrix}}\right]}$, and taking the Hadamard product of the vectors (entrywise product), denoted ${\displaystyle d\odot v}$:

${\displaystyle Dv=d\odot v={\begin{bmatrix}a_{1}\\\vdots \\a_{n}\end{bmatrix}}\odot {\begin{bmatrix}x_{1}\\\vdots \\x_{n}\end{bmatrix}}={\begin{bmatrix}a_{1}x_{1}\\\vdots \\a_{n}x_{n}\end{bmatrix}}.}$

This is mathematically equivalent, but avoids storing all the zero terms of this sparse matrix. This product is thus used in computational mathematics, such as computing products of derivatives in backpropagation, since is these cases the entire diagonal matrix is not stored.

## Matrix operations

The operations of matrix addition and matrix multiplication are especially simple for diagonal matrices. Write for a diagonal matrix whose diagonal entries starting in the upper left corner are a1, ..., an. Then, for addition, we have

+ =

and for matrix multiplication,

· = .

The diagonal matrix is invertible if and only if the entries a1, ..., an are all non-zero. In this case, we have

= .

In particular, the diagonal matrices form a subring of the ring of all n-by-n matrices.

Multiplying an n-by-n matrix A from the left with amounts to multiplying the ith row of A by ai for all i; multiplying the matrix A from the right with amounts to multiplying the ith column of A by ai for all i.

## Operator matrix in eigenbasis

As explained in determining coefficients of operator matrix, there is a special basis, e1, ..., en, for which the matrix ${\displaystyle A}$ takes the diagonal form. Hence, in the defining equation ${\displaystyle A{\vec {e}}_{j}=\sum a_{i,j}{\vec {e}}_{i}}$, all coefficients ${\displaystyle a_{i,j}}$ with i ? j are zero, leaving only one term per sum. The surviving diagonal elements, ${\displaystyle a_{i,i}}$, are known as eigenvalues and designated with ${\displaystyle \lambda _{i}}$ in the equation, which reduces to ${\displaystyle A{\vec {e}}_{i}=\lambda _{i}{\vec {e}}_{i}}$. The resulting equation is known as eigenvalue equation[2] and used to derive the characteristic polynomial and, further, eigenvalues and eigenvectors.

In other words, the eigenvalues of are ?1, ..., ?n with associated eigenvectors of e1, ..., en.

## Properties

The determinant of is the product a1...an.

The adjugate of a diagonal matrix is again diagonal.

A square matrix is diagonal if and only if it is triangular and normal.

Any square diagonal matrix is also a symmetric matrix.

A symmetric diagonal matrix can be defined as a matrix that is both upper- and lower-triangular. The identity matrix In and any square zero matrix are diagonal. A one-dimensional matrix is always diagonal.

## Applications

Diagonal matrices occur in many areas of linear algebra. Because of the simple description of the matrix operation and eigenvalues/eigenvectors given above, it is typically desirable to represent a given matrix or linear map by a diagonal matrix.

In fact, a given n-by-n matrix A is similar to a diagonal matrix (meaning that there is a matrix X such that X-1AX is diagonal) if and only if it has n linearly independent eigenvectors. Such matrices are said to be diagonalizable.

Over the field of real or complex numbers, more is true. The spectral theorem says that every normal matrix is unitarily similar to a diagonal matrix (if AA* = A*A then there exists a unitary matrix U such that UAU* is diagonal). Furthermore, the singular value decomposition implies that for any matrix A, there exist unitary matrices U and V such that UAV* is diagonal with positive entries.

## Operator theory

In operator theory, particularly the study of PDEs, operators are particularly easy to understand and PDEs easy to solve if the operator is diagonal with respect to the basis with which one is working; this corresponds to a separable partial differential equation. Therefore, a key technique to understanding operators is a change of coordinates--in the language of operators, an integral transform--which changes the basis to an eigenbasis of eigenfunctions: which makes the equation separable. An important example of this is the Fourier transform, which diagonalizes constant coefficient differentiation operators (or more generally translation invariant operators), such as the Laplacian operator, say, in the heat equation.

Especially easy are multiplication operators, which are defined as multiplication by (the values of) a fixed function-the values of the function at each point correspond to the diagonal entries of a matrix.

## Notes

1. ^ Proof: given the elementary matrix ${\displaystyle e_{ij}}$, ${\displaystyle Me_{ij}}$ is the matrix with only the i-th row of M and ${\displaystyle e_{ij}M}$ is the square matrix with only the M j-th column, so the non-diagonal entries must be zero, and the ith diagonal entry much equal the jth diagonal entry.
2. ^ Over more general rings, this does not hold, because one cannot always divide.

## References

1. ^ "Do Diagonal Matrices Always Commute?". Stack Exchange. March 15, 2016. Retrieved 2018.
2. ^ Nearing, James (2010). "Chapter 7.9: Eigenvalues and Eigenvectors" (PDF). Mathematical Tools for Physics. ISBN 048648212X. Retrieved 2012.
• Horn, Roger A.; Johnson, Charles R. (1985). Matrix Analysis. Cambridge University Press. ISBN 0-521-30586-1.

This article uses material from the Wikipedia page available here. It is released under the Creative Commons Attribution-Share-Alike License 3.0.