Visual Tools
Calculators
Tables
Mathematical Keyboard
Converters
Other Tools


Characteristic Equation






The Polynomial Whose Roots Are the Eigenvalues

The eigenvalue problem Av = λv converts into a determinant condition: det(A − λI) = 0. This determinant is a polynomial in λ whose roots are the eigenvalues. Computing the characteristic polynomial and factoring it is the standard method for finding eigenvalues of small matrices — and the polynomial's coefficients encode the trace, determinant, and other invariants of the matrix.



From Eigenvectors to the Determinant Condition

The equation Av=λvA\mathbf{v} = \lambda\mathbf{v} rearranges to (AλI)v=0(A - \lambda I)\mathbf{v} = \mathbf{0}. This is a homogeneous system, and eigenvectors are its nontrivial solutions. Nontrivial solutions exist if and only if the coefficient matrix AλIA - \lambda I is singular:

det(AλI)=0\det(A - \lambda I) = 0


This is the characteristic equation. It holds for exactly those values of λ\lambda that are eigenvalues of AA. Every other value of λ\lambda makes AλIA - \lambda I invertible, the system has only the trivial solution, and no eigenvector exists for that λ\lambda.

The characteristic equation transforms the geometric question "which directions does AA preserve?" into the algebraic question "for which λ\lambda is this determinant zero?"

The Characteristic Polynomial

The expression p(λ)=det(AλI)p(\lambda) = \det(A - \lambda I) is a polynomial of degree nn in the variable λ\lambda. It is called the characteristic polynomial of AA.

For an n×nn \times n matrix, p(λ)p(\lambda) has degree nn with leading term (1)nλn(-1)^n \lambda^n. The constant term is p(0)=det(A)p(0) = \det(A) — the determinant of the matrix itself. The coefficient of λn1\lambda^{n-1} is (1)n1tr(A)(-1)^{n-1}\text{tr}(A), connecting the next-to-leading term to the trace.

The eigenvalues are precisely the roots of p(λ)=0p(\lambda) = 0. Every root is an eigenvalue, and every eigenvalue is a root. The characteristic polynomial packages the entire eigenvalue structure of the matrix into a single algebraic expression.

Computing the Characteristic Polynomial: 2×2

For A=(abcd)A = \begin{pmatrix} a & b \\ c & d \end{pmatrix}, the characteristic polynomial is

p(λ)=det(aλbcdλ)=(aλ)(dλ)bc=λ2(a+d)λ+(adbc)p(\lambda) = \det\begin{pmatrix} a - \lambda & b \\ c & d - \lambda \end{pmatrix} = (a - \lambda)(d - \lambda) - bc = \lambda^2 - (a + d)\lambda + (ad - bc)


This is λ2tr(A)λ+det(A)\lambda^2 - \text{tr}(A)\lambda + \det(A). The eigenvalues follow from the quadratic formula:

λ=tr(A)±tr(A)24det(A)2\lambda = \frac{\text{tr}(A) \pm \sqrt{\text{tr}(A)^2 - 4\det(A)}}{2}


The discriminant Δ=tr(A)24det(A)\Delta = \text{tr}(A)^2 - 4\det(A) classifies the eigenvalue type. When Δ>0\Delta > 0, there are two distinct real eigenvalues. When Δ=0\Delta = 0, there is one repeated real eigenvalue. When Δ<0\Delta < 0, the eigenvalues are a complex conjugate pair.

Worked Example


For A=(5234)A = \begin{pmatrix} 5 & 2 \\ 3 & 4 \end{pmatrix}: tr(A)=9\text{tr}(A) = 9, det(A)=14\det(A) = 14, Δ=8156=25\Delta = 81 - 56 = 25. The eigenvalues are λ=9±52\lambda = \frac{9 \pm 5}{2}, giving λ1=7\lambda_1 = 7 and λ2=2\lambda_2 = 2.

Computing the Characteristic Polynomial: 3×3

For a 3×33 \times 3 matrix, expanding det(AλI)\det(A - \lambda I) using cofactor expansion produces a cubic polynomial:

p(λ)=λ3+tr(A)λ2(sum of 2×2 principal minors)λ+det(A)p(\lambda) = -\lambda^3 + \text{tr}(A)\lambda^2 - (\text{sum of } 2 \times 2 \text{ principal minors})\lambda + \det(A)


The computation is lengthier but follows the same cofactor mechanics as any 3×33 \times 3 determinant.

Worked Example


For A=(210031001)A = \begin{pmatrix} 2 & 1 & 0 \\ 0 & 3 & 1 \\ 0 & 0 & 1 \end{pmatrix}, this is upper triangular, so AλIA - \lambda I is also upper triangular with diagonal entries 2λ2 - \lambda, 3λ3 - \lambda, 1λ1 - \lambda:

p(λ)=(2λ)(3λ)(1λ)p(\lambda) = (2 - \lambda)(3 - \lambda)(1 - \lambda)


The eigenvalues are λ=1,2,3\lambda = 1, 2, 3 — readable directly from the diagonal. For triangular matrices, the characteristic polynomial always factors as the product of the diagonal terms, making the eigenvalues visible by inspection.

For non-triangular 3×33 \times 3 matrices, the cubic must be factored by finding rational roots (testing factors of the constant term), by inspection, or by the cubic formula.

Larger Matrices

For an n×nn \times n matrix, the characteristic polynomial has degree nn, and finding its roots becomes increasingly difficult as nn grows. There is no general closed-form formula for roots of polynomials of degree 55 or higher (Abel-Ruffini theorem), so explicit factoring is limited to small matrices or matrices with special structure.

Diagonal and triangular matrices are immediate: the eigenvalues are the diagonal entries. Block triangular matrices factor block by block: the characteristic polynomial is the product of the characteristic polynomials of the diagonal blocks.

For general large matrices, eigenvalues are computed numerically by iterative algorithms — most importantly the QR algorithm, which repeatedly applies QR decompositions to converge on the eigenvalues without ever forming the characteristic polynomial explicitly. Computing the polynomial and then finding its roots is numerically unstable for large nn and is never used in practice.

Algebraic Multiplicity

If λ0\lambda_0 is a root of the characteristic polynomial p(λ)p(\lambda), its algebraic multiplicity is the largest power kk such that (λλ0)k(\lambda - \lambda_0)^k divides p(λ)p(\lambda). Equivalently, it is the multiplicity of λ0\lambda_0 as a root.

If p(λ)=(λ2)3(λ+1)p(\lambda) = (\lambda - 2)^3(\lambda + 1), then λ=2\lambda = 2 has algebraic multiplicity 33 and λ=1\lambda = -1 has algebraic multiplicity 11. The algebraic multiplicities of all eigenvalues sum to nn — the degree of the polynomial — when complex roots are included.

The algebraic multiplicity is an upper bound for the geometric multiplicity: 1mg(λ)ma(λ)1 \leq m_g(\lambda) \leq m_a(\lambda). The geometric multiplicity is the dimension of the eigenspace, and it can be strictly smaller than the algebraic multiplicity. When this gap occurs for any eigenvalue, the matrix is not diagonalizable.

Finding Eigenvectors After Finding Eigenvalues

Once the eigenvalues are known, the eigenvectors for each λi\lambda_i are found by solving the homogeneous system (AλiI)v=0(A - \lambda_i I)\mathbf{v} = \mathbf{0}.

Row reduce AλiIA - \lambda_i I and express the general solution in parametric form. Each free variable contributes one basis vector for the eigenspace EλiE_{\lambda_i}.

Worked Example


For A=(1243)A = \begin{pmatrix} 1 & 2 \\ 4 & 3 \end{pmatrix}, the characteristic polynomial is λ24λ5=(λ5)(λ+1)\lambda^2 - 4\lambda - 5 = (\lambda - 5)(\lambda + 1). Eigenvalues: λ1=5\lambda_1 = 5, λ2=1\lambda_2 = -1.

For λ1=5\lambda_1 = 5: A5I=(4242)A - 5I = \begin{pmatrix} -4 & 2 \\ 4 & -2 \end{pmatrix}. Row reducing: (11/200)\begin{pmatrix} 1 & -1/2 \\ 0 & 0 \end{pmatrix}. Free variable v2=tv_2 = t, so v1=t/2v_1 = t/2. Eigenvector: v1=(1,2)T\mathbf{v}_1 = (1, 2)^T.

For λ2=1\lambda_2 = -1: A+I=(2244)A + I = \begin{pmatrix} 2 & 2 \\ 4 & 4 \end{pmatrix}. Row reducing: (1100)\begin{pmatrix} 1 & 1 \\ 0 & 0 \end{pmatrix}. Free variable v2=tv_2 = t, so v1=tv_1 = -t. Eigenvector: v2=(1,1)T\mathbf{v}_2 = (-1, 1)^T.

Verification: Av1=(510)=5v1A\mathbf{v}_1 = \begin{pmatrix} 5 \\ 10 \end{pmatrix} = 5\mathbf{v}_1 and Av2=(11)=1v2A\mathbf{v}_2 = \begin{pmatrix} 1 \\ -1 \end{pmatrix} = -1 \cdot \mathbf{v}_2.

The Cayley-Hamilton Theorem

Every square matrix satisfies its own characteristic polynomial. If p(λ)=det(AλI)p(\lambda) = \det(A - \lambda I) is the characteristic polynomial, then

p(A)=0p(A) = 0


where 00 is the zero matrix and λ\lambda is replaced by AA (with constant terms multiplied by II).

For example, if p(λ)=λ25λ+6p(\lambda) = \lambda^2 - 5\lambda + 6, then A25A+6I=OA^2 - 5A + 6I = O. This can be rearranged to express A1A^{-1} as a polynomial in AA: A1=16(5IA)A^{-1} = \frac{1}{6}(5I - A) (provided det(A)=60\det(A) = 6 \neq 0). More generally, the Cayley-Hamilton theorem guarantees that A1A^{-1} can always be written as a polynomial in AA of degree at most n1n - 1.

The theorem also shows that any power AkA^k with knk \geq n can be reduced to a polynomial in AA of degree at most n1n - 1 — the characteristic polynomial provides a recurrence that expresses higher powers in terms of lower ones.

Characteristic Polynomial and Similarity

Similar matrices have the same characteristic polynomial:

det(P1APλI)=det(P1(AλI)P)=det(AλI)\det(P^{-1}AP - \lambda I) = \det(P^{-1}(A - \lambda I)P) = \det(A - \lambda I)


The second equality uses the multiplicative property of the determinant: det(P1)det(AλI)det(P)=det(AλI)\det(P^{-1})\det(A - \lambda I)\det(P) = \det(A - \lambda I), since det(P1)det(P)=1\det(P^{-1})\det(P) = 1.

This means the characteristic polynomial is a property of the linear transformation itself, not of any particular matrix representation. Changing the basis changes the matrix but not the polynomial. Since the eigenvalues are the roots of the polynomial, similar matrices have the same eigenvalues with the same algebraic multiplicities.

The trace and determinant are just two of the nn coefficients of the characteristic polynomial. The polynomial carries more information than either one alone — it determines the complete multiset of eigenvalues, not just their sum and product.