Linear Algebra 63 | Spectral Mapping Theorem
Based on The Bright Side of Mathematics's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Eigenvalues of A^m follow the rule: if A x = λ x, then A^m x = λ^m x for all natural m.
Briefing
The spectral mapping theorem for polynomials gives a direct rule for how eigenvalues change when a matrix is transformed by a polynomial: the spectrum of p(A) is exactly the set of values p(λ) as λ ranges over the spectrum of A. That matters because it turns what could be a fresh eigenvalue computation for a new matrix into a simple “plug-in” operation on the old eigenvalues.
The discussion starts from the eigenvalue equation A x = Λ x, emphasizing complex-valued matrices so eigenvalues may be complex as well. Multiplying by A again shows that A^2 x = Λ^2 x, so x remains an eigenvector for A^2 while the eigenvalue squares. Repeating the argument (and using induction) yields the general pattern: for every natural power m, A^m has the same eigenvectors as A, with eigenvalues Λ^m. This power behavior extends naturally from monomials to polynomials.
A polynomial p(z) = c_m z^m + … + c_1 z + c_0 is turned into a matrix polynomial p(A) by substituting A for z. The constant term requires the identity matrix: c_0 I, so the whole expression becomes a well-defined n×n matrix. With that setup, the spectral mapping theorem can be stated: the spectrum of p(A) is determined completely by the spectrum of A, and specifically equals {p(Λ) : Λ ∈ spectrum(A)}. As a consequence, if spectrum(A) has n elements, then spectrum(p(A)) has at most n elements. A constant polynomial illustrates the extreme case: if p(z)=c_0, then p(A)=c_0 I and the spectrum collapses to the single value {c_0}.
The proof strategy is framed as showing equality of two sets via two inclusions. One inclusion follows from the earlier eigenvalue-power reasoning (with matrix additions handled by induction). The other inclusion is proved by contrapositive. First, the constant-polynomial case is handled directly using the characteristic polynomial: eigenvalues correspond to zeros of det(p(A) − λ̃ I). For non-constant p, the argument assumes a complex number μ is not in {p(Λ)} and builds a new polynomial q(z)=p(z)−μ. Factoring q using the fundamental theorem of algebra gives q(z)=c ∏_{j=1}^m (z−A_j). Since μ was chosen to avoid p(Λ) for all eigenvalues Λ of A, none of the roots A_j can lie in spectrum(A), meaning each factor det(A − A_j I) is nonzero. Multiplicativity of determinants then shows det(q(A)) ≠ 0, so μ cannot be an eigenvalue of p(A). That completes the contrapositive and therefore the set equality.
A concrete example closes the loop. For A = [[3,2],[1,2]], the eigenvalues are 1 and 4. Taking B = 3A^3 − 7A^2 + A − 2I, the theorem says the eigenvalues of B are obtained by evaluating the polynomial at 1 and 4. The results are −5 and 82, avoiding any direct eigenvalue computation for B itself.
Cornell Notes
The spectral mapping theorem for polynomials links eigenvalues before and after a polynomial matrix transformation. If A has spectrum spectrum(A), then for any polynomial p, the spectrum of p(A) is exactly {p(λ) : λ ∈ spectrum(A)}. The key mechanism comes from the eigenvalue equation: if A x = λ x, then A^m x = λ^m x for all natural m, and this extends from powers to general polynomials. The theorem implies eigenvalues of p(A) can be found by plugging the eigenvalues of A into p, often eliminating a new characteristic-polynomial calculation. A proof by contrapositive uses factorization of q(z)=p(z)−μ and determinant multiplicativity to show that any μ not of the form p(λ) cannot be an eigenvalue of p(A).
Why does an eigenvector of A stay an eigenvector for A^m, and how do the eigenvalues change?
How is a polynomial p(z) turned into a matrix p(A), and why does the identity matrix appear?
What does the spectral mapping theorem claim about the spectrum of p(A)?
How does the contrapositive proof work for non-constant polynomials?
In the example, how are the eigenvalues of B obtained from those of A?
Review Questions
- If A x = λ x, what is A^5 x and what eigenvalue does A^5 have on the vector x?
- Given a polynomial p(z)=z^2−3z+2 and a matrix A with eigenvalues 1 and 4, what set does the spectral mapping theorem predict for spectrum(p(A))?
- Why must the constant term in p(A) be written as c_0 I rather than just c_0?
Key Points
- 1
Eigenvalues of A^m follow the rule: if A x = λ x, then A^m x = λ^m x for all natural m.
- 2
A polynomial matrix p(A) is formed by substituting A for the variable and using c_0 I for the constant term.
- 3
For any polynomial p, spectrum(p(A)) equals {p(λ) : λ ∈ spectrum(A)}.
- 4
The theorem implies spectrum(p(A)) has at most as many elements as spectrum(A), and constant polynomials collapse the spectrum to a single value.
- 5
A contrapositive proof for non-constant p uses q(z)=p(z)−μ, factorization into linear terms, and determinant multiplicativity to show μ cannot be an eigenvalue unless μ = p(λ) for some eigenvalue λ of A.
- 6
In the worked example, evaluating the polynomial at A’s eigenvalues (1 and 4) immediately yields B’s eigenvalues (−5 and 82).