The eigenvalue problem Av = λv converts into a determinant condition: det(A − λI) = 0. This determinant is a polynomial in λ whose roots are the eigenvalues. Computing the characteristic polynomial and factoring it is the standard method for finding eigenvalues of small matrices — and the polynomial's coefficients encode the trace, determinant, and other invariants of the matrix.
From Eigenvectors to the Determinant Condition
The equation Av=λv rearranges to (A−λI)v=0. This is a homogeneous system, and eigenvectors are its nontrivial solutions. Nontrivial solutions exist if and only if the coefficient matrix A−λI is singular:
det(A−λI)=0
This is the characteristic equation. It holds for exactly those values of λ that are eigenvalues of A. Every other value of λ makes A−λIinvertible, the system has only the trivial solution, and no eigenvector exists for that λ.
The characteristic equation transforms the geometric question "which directions does A preserve?" into the algebraic question "for which λ is this determinant zero?"
The Characteristic Polynomial
The expression p(λ)=det(A−λI) is a polynomial of degree n in the variable λ. It is called the characteristic polynomial of A.
For an n×n matrix, p(λ) has degree n with leading term (−1)nλn. The constant term is p(0)=det(A) — the determinant of the matrix itself. The coefficient of λn−1 is (−1)n−1tr(A), connecting the next-to-leading term to the trace.
The eigenvalues are precisely the roots of p(λ)=0. Every root is an eigenvalue, and every eigenvalue is a root. The characteristic polynomial packages the entire eigenvalue structure of the matrix into a single algebraic expression.
This is λ2−tr(A)λ+det(A). The eigenvalues follow from the quadratic formula:
λ=2tr(A)±tr(A)2−4det(A)
The discriminant Δ=tr(A)2−4det(A) classifies the eigenvalue type. When Δ>0, there are two distinct real eigenvalues. When Δ=0, there is one repeated real eigenvalue. When Δ<0, the eigenvalues are a complex conjugate pair.
Worked Example
For A=(5324): tr(A)=9, det(A)=14, Δ=81−56=25. The eigenvalues are λ=29±5, giving λ1=7 and λ2=2.
Computing the Characteristic Polynomial: 3×3
For a 3×3 matrix, expanding det(A−λI) using cofactor expansion produces a cubic polynomial:
p(λ)=−λ3+tr(A)λ2−(sum of 2×2 principal minors)λ+det(A)
The computation is lengthier but follows the same cofactor mechanics as any 3×3 determinant.
Worked Example
For A=200130011, this is upper triangular, so A−λI is also upper triangular with diagonal entries 2−λ, 3−λ, 1−λ:
p(λ)=(2−λ)(3−λ)(1−λ)
The eigenvalues are λ=1,2,3 — readable directly from the diagonal. For triangular matrices, the characteristic polynomial always factors as the product of the diagonal terms, making the eigenvalues visible by inspection.
For non-triangular 3×3 matrices, the cubic must be factored by finding rational roots (testing factors of the constant term), by inspection, or by the cubic formula.
Larger Matrices
For an n×n matrix, the characteristic polynomial has degree n, and finding its roots becomes increasingly difficult as n grows. There is no general closed-form formula for roots of polynomials of degree 5 or higher (Abel-Ruffini theorem), so explicit factoring is limited to small matrices or matrices with special structure.
Diagonal and triangular matrices are immediate: the eigenvalues are the diagonal entries. Block triangular matrices factor block by block: the characteristic polynomial is the product of the characteristic polynomials of the diagonal blocks.
For general large matrices, eigenvalues are computed numerically by iterative algorithms — most importantly the QR algorithm, which repeatedly applies QR decompositions to converge on the eigenvalues without ever forming the characteristic polynomial explicitly. Computing the polynomial and then finding its roots is numerically unstable for large n and is never used in practice.
Algebraic Multiplicity
If λ0 is a root of the characteristic polynomial p(λ), its algebraic multiplicity is the largest power k such that (λ−λ0)k divides p(λ). Equivalently, it is the multiplicity of λ0 as a root.
If p(λ)=(λ−2)3(λ+1), then λ=2 has algebraic multiplicity 3 and λ=−1 has algebraic multiplicity 1. The algebraic multiplicities of all eigenvalues sum to n — the degree of the polynomial — when complex roots are included.
The algebraic multiplicity is an upper bound for the geometric multiplicity: 1≤mg(λ)≤ma(λ). The geometric multiplicity is the dimension of the eigenspace, and it can be strictly smaller than the algebraic multiplicity. When this gap occurs for any eigenvalue, the matrix is not diagonalizable.
Finding Eigenvectors After Finding Eigenvalues
Once the eigenvalues are known, the eigenvectors for each λi are found by solving the homogeneous system (A−λiI)v=0.
Row reduce A−λiI and express the general solution in parametric form. Each free variable contributes one basis vector for the eigenspace Eλi.
Worked Example
For A=(1423), the characteristic polynomial is λ2−4λ−5=(λ−5)(λ+1). Eigenvalues: λ1=5, λ2=−1.
For λ1=5: A−5I=(−442−2). Row reducing: (10−1/20). Free variable v2=t, so v1=t/2. Eigenvector: v1=(1,2)T.
For λ2=−1: A+I=(2424). Row reducing: (1010). Free variable v2=t, so v1=−t. Eigenvector: v2=(−1,1)T.
Verification: Av1=(510)=5v1 and Av2=(1−1)=−1⋅v2.
The Cayley-Hamilton Theorem
Every square matrix satisfies its own characteristic polynomial. If p(λ)=det(A−λI) is the characteristic polynomial, then
p(A)=0
where 0 is the zero matrix and λ is replaced by A (with constant terms multiplied by I).
For example, if p(λ)=λ2−5λ+6, then A2−5A+6I=O. This can be rearranged to express A−1 as a polynomial in A: A−1=61(5I−A) (provided det(A)=6=0). More generally, the Cayley-Hamilton theorem guarantees that A−1 can always be written as a polynomial in A of degree at most n−1.
The theorem also shows that any power Ak with k≥n can be reduced to a polynomial in A of degree at most n−1 — the characteristic polynomial provides a recurrence that expresses higher powers in terms of lower ones.
Characteristic Polynomial and Similarity
Similar matrices have the same characteristic polynomial:
det(P−1AP−λI)=det(P−1(A−λI)P)=det(A−λI)
The second equality uses the multiplicative property of the determinant: det(P−1)det(A−λI)det(P)=det(A−λI), since det(P−1)det(P)=1.
This means the characteristic polynomial is a property of the linear transformation itself, not of any particular matrix representation. Changing the basis changes the matrix but not the polynomial. Since the eigenvalues are the roots of the polynomial, similar matrices have the same eigenvalues with the same algebraic multiplicities.
The trace and determinant are just two of the n coefficients of the characteristic polynomial. The polynomial carries more information than either one alone — it determines the complete multiset of eigenvalues, not just their sum and product.