Identity matrix
In linear algebra, the identity matrix is a square matrix containing ones along its main diagonal and zeros in all other positions, serving as the multiplicative identity element for matrix multiplication.[1] Denoted by I_n where n specifies the matrix dimension, it satisfies I_n \mathbf{x} = \mathbf{x} for any compatible vector \mathbf{x}, effectively leaving vectors unchanged under linear transformation.[2] For example, the 2×2 identity matrix isand multiplying it by any 2×2 matrix A yields A itself.[3] A fundamental property of the identity matrix is its role as the unit element in the ring of square matrices: for any n \times n matrix A, A I_n = I_n A = A, ensuring it commutes with all such matrices and acts as a neutral operation in algebraic structures.[4] This property extends to its use in defining matrix inverses, where a matrix A is invertible if there exists another matrix B such that A B = B A = I_n, highlighting the identity's centrality in solvability and decomposition of linear systems.[5] Additionally, the identity matrix is symmetric, orthogonal (for real entries), and unitary (for complex entries), preserving norms and inner products in vector spaces.[6] Beyond algebra, the identity matrix represents the trivial linear transformation that maps every vector to itself, making it indispensable in applications such as solving differential equations, computer graphics for no-op transformations, and numerical methods where it initializes or benchmarks matrix operations.[7] Its simplicity belies its ubiquity, as it appears in eigenvalue decompositions (with eigenvalue 1 for all eigenvectors) and as a building block for permutation and diagonal matrices.[1]$$ I_2 = \begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix} $$$$ I_2 = \begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix} $$
Definition and Notation
Definition
In linear algebra, the identity matrix of order n, where n is a positive integer, is defined as the n \times n square matrix whose main diagonal entries are all equal to 1 and all other entries are equal to 0.[8] This structure ensures that the matrix has exactly n rows and n columns, with the diagonal running from the top-left to the bottom-right.[9] For illustration, the identity matrix of order 2 takes the form \begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix}, while the identity matrix of order 3 is \begin{pmatrix} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end{pmatrix}. [8] These examples highlight the pattern: ones only on the principal diagonal and zeros off-diagonal. The entries of the identity matrix can be precisely expressed using the Kronecker delta symbol \delta_{ij}, defined as \delta_{ij} = 1 if i = j and \delta_{ij} = 0 otherwise, for indices i, j = 1, 2, \dots, n. Thus, the (i,j)-th entry of the identity matrix is \delta_{ij}, which compactly captures its diagonal structure.[10] The identity matrix must be square, as non-square analogs do not exist in standard linear algebra, where the concept is tied to the dimensions required for multiplicative identity properties in square matrix multiplication.[11]Notation and Terminology
The identity matrix is most commonly denoted by the bold capital letter \mathbf{I} when the dimension is clear from context, or by \mathbf{I}_n to explicitly indicate an n \times n square matrix.[1][2] Alternative notations include E or E_n, derived from the German term "Einheitsmatrix" for unit matrix, particularly in older European mathematical literature; the subscript n is consistently used to specify the matrix size across these variants.[1] Standard terminology refers to it as the "identity matrix," reflecting its role as the multiplicative identity in matrix algebra, though it is also known as the "unit matrix," a term borrowed from the analogous "unit element" in number theory and ring theory.[1] It may additionally be described as a "principal diagonal matrix with ones," emphasizing its diagonal structure of all 1s and off-diagonal zeros.[12] The notation and terminology remain unchanged when the identity matrix is defined over the complex numbers or other fields, with the real numbers serving as the default context unless specified otherwise.[1]Properties
Algebraic Properties
The identity matrix I_n acts as the multiplicative identity element in the algebra of square matrices over the real or complex numbers. For any m \times n matrix A, the product A I_n equals A, and similarly I_m A = A, where I_m is the m \times m identity matrix.[4] This property follows from the definition of matrix multiplication: the (i,j)-th entry of A I_n is \sum_{k=1}^n a_{ik} (I_n)_{kj} = \sum_{k=1}^n a_{ik} \delta_{kj} = a_{ij}, where \delta_{kj} is the Kronecker delta (equal to 1 if k = j and 0 otherwise). The same reasoning applies to I_m A by symmetry in the multiplication rule.[2] To illustrate, consider a $2 \times 2 matrix A = \begin{pmatrix} a & b \\ c & d \end{pmatrix} and I_2 = \begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix}. The product is A I_2 = \begin{pmatrix} a & b \\ c & d \end{pmatrix} \begin{pmatrix} 1 & 0 \\ 0 & 1 \end{pmatrix} = \begin{pmatrix} a \cdot 1 + b \cdot 0 & a \cdot 0 + b \cdot 1 \\ c \cdot 1 + d \cdot 0 & c \cdot 0 + d \cdot 1 \end{pmatrix} = \begin{pmatrix} a & b \\ c & d \end{pmatrix} = A. A parallel computation shows I_2 A = A.[3] The identity matrix is self-inverse, meaning I_n^{-1} = I_n, since I_n I_n = I_n. This directly stems from its role as the multiplicative identity.[1] For square matrices A of order n, the identity commutes with every element: I_n A = A I_n = A. This commutativity holds because both products reduce to A via the multiplicative identity property.[4] Additively, I_n + O_n = I_n, where O_n is the n \times n zero matrix, but I_n is not the additive identity (which is O_n). The trace of I_n equals n, as it sums the diagonal entries of 1. The determinant of I_n is 1, reflecting its unimodular nature in matrix groups.[6]Analytic and Structural Properties
The identity matrix I_n of order n has all eigenvalues equal to 1, with algebraic multiplicity n.[13] Every nonzero vector in \mathbb{R}^n (or the appropriate vector space) serves as an eigenvector corresponding to this eigenvalue, since I_n \mathbf{v} = \mathbf{v} for any \mathbf{v}.[13] In particular, the standard basis vectors e_i (with 1 in the i-th position and zeros elsewhere) form a set of n linearly independent eigenvectors.[12] The characteristic polynomial of I_n is \det(I_n - \lambda I_n) = (1 - \lambda)^n.[14] This monic polynomial of degree n reflects the repeated root at \lambda = 1, confirming the eigenvalue structure.[14] As a diagonal matrix with 1's on the main diagonal, I_n is already in diagonal form, requiring no further diagonalization.[15] It is orthogonal, satisfying I_n^T I_n = I_n and I_n^T = I_n, which preserves the Euclidean norm under multiplication.[15] For real entries, I_n is symmetric and positive definite, as the quadratic form \mathbf{x}^T I_n \mathbf{x} = \|\mathbf{x}\|^2 > 0 for all nonzero \mathbf{x} \in \mathbb{R}^n.[16] The rank of I_n is n, equal to its dimension, indicating full column and row rank.[17] The null space (kernel) of I_n is the trivial subspace \{\mathbf{0}\}, since I_n \mathbf{x} = \mathbf{0} implies \mathbf{x} = \mathbf{0}./07:_Spectral_Theory/7.01:_Eigenvalues_and_Eigenvectors_of_a_Matrix) In power series expansions, the powers satisfy I_n^k = I_n for all integers k \geq 1, leading to idempotence under multiplication.[18] The matrix exponential is e^{t I_n} = e^t I_n, derived from the series \sum_{k=0}^\infty \frac{(t I_n)^k}{k!} = I_n \sum_{k=0}^\infty \frac{t^k}{k!} = e^t I_n.[19] The identity matrix is unique in the algebra of n \times n matrices over a field, as it is the only matrix E satisfying A E = E A = A for every matrix A.[18] This uniqueness underscores its role as the multiplicative identity element.[18]Relations and Applications
Relations to Other Concepts
The identity matrix serves as a special case of a diagonal matrix, where all diagonal entries are equal to 1 and all off-diagonal entries are 0, distinguishing it from general diagonal matrices that may have arbitrary nonzero values on the diagonal while maintaining zeros elsewhere.[20][21] In contrast, a general diagonal matrix scales vectors along the coordinate axes by its diagonal elements, whereas the identity matrix leaves them unchanged.[22] The identity matrix is also the permutation matrix corresponding to the identity permutation, obtained by rearranging no rows or columns of itself, which sets it apart from other permutation matrices that reorder entries to represent nontrivial permutations.[23][24] Permutation matrices preserve the structure of linear transformations as bijections on basis vectors, but only the identity permutation yields the identity matrix itself.[25] In the context of tensor products, the Kronecker product of two identity matrices satisfies I_m \otimes I_n = I_{mn}, where I_k denotes the k \times k identity matrix, illustrating how the identity extends naturally to higher-dimensional structures without alteration.[26][27] This property underscores the identity matrix's role as a multiplicative neutral element in the algebra of Kronecker products.[28] The identity matrix is orthogonal, satisfying I^T I = I, and it has operator norm 1 with respect to the Euclidean vector norm, as it preserves lengths without scaling or rotation.[15][29] Unlike general orthogonal matrices, which represent isometries such as rotations or reflections, the identity matrix corresponds to the trivial isometry that fixes every vector.[30] In comparison to the zero matrix, which acts as the additive identity in the ring of matrices under addition—satisfying A + 0 = A for any matrix A—the identity matrix is the multiplicative identity under matrix multiplication, where A I = I A = A for compatible square matrices A.[31][32] These roles do not overlap, as the zero matrix annihilates under multiplication while the identity preserves under both operations in their respective contexts.[33] Abstractly, the identity matrix represents the identity element in the monoid of n \times n square matrices over a field, equipped with matrix multiplication as the operation, where every matrix has a unique right and left identity given by I_n.[34][35] This monoidal structure highlights the identity matrix's foundational position in algebraic frameworks for linear transformations.[36]Applications in Mathematics and Beyond
In linear algebra, the identity matrix plays a crucial role in solving systems of linear equations, particularly when computing matrix inverses via Gaussian elimination. To find the inverse of an invertible square matrix A, one augments A with the identity matrix of the same dimension to form the matrix [A \mid I], then applies row operations to transform the left block into the identity matrix; the right block then becomes A^{-1}.[37] This method leverages the identity's property as the neutral element under matrix multiplication, ensuring the transformations preserve the solution.[38] The identity matrix also appears in change-of-basis transformations, where it signifies no alteration in the coordinate system. In the similarity transformation P^{-1} A P = B, setting P = I yields B = A, illustrating that the identity matrix leaves the matrix representation unchanged under the standard basis.[39] This is fundamental in preserving invariants like eigenvalues across equivalent bases.[40] In computational mathematics and software libraries, the identity matrix serves as a default or initializing structure for algorithms and data handling. For instance, the NumPy library in Python provides thenumpy.eye(n) function to generate an n \times n identity matrix, which is commonly used to initialize transformation matrices or as a starting point in iterative solvers. Additionally, it acts as a test case for verifying matrix function algorithms, such as those computing exponentials or logarithms, by exploiting functional identities like f(I) = I for certain analytic functions f, allowing assessment of backward stability through residual bounds.
In physics, the identity matrix represents the identity transformation, corresponding to scenarios with no external influences. In quantum mechanics, the identity operator \hat{I}, whose matrix representation is the identity matrix in any orthonormal basis, leaves quantum states unchanged: \hat{I} |\psi\rangle = |\psi\rangle for any state vector |\psi\rangle, serving as the resolution of the identity in spectral decompositions.[41] In classical mechanics, the identity transformation arises in canonical formulations for free particles under no forces, where the Hamiltonian remains unaltered, mapping coordinates and momenta identically as in the absence of interactions.[42]
In statistics, the identity matrix models the covariance structure of uncorrelated random variables. For a multivariate distribution with uncorrelated components, each scaled by a variance \sigma^2, the covariance matrix is \sigma^2 I, indicating zero covariances off the diagonal and equal variances on the diagonal; this simplifies computations in models like the multivariate normal, where it implies independence under Gaussian assumptions.[43][44]
In graph theory, the identity matrix represents the adjacency matrix of a graph consisting of isolated vertices each with a self-loop. Unlike the zero matrix, which corresponds to an empty graph with no edges or loops, the identity matrix encodes a structure where each vertex connects only to itself, facilitating analysis of walks that remain at the same vertex.[45] This configuration is useful in studying graph spectra and connectivity properties.[46]