Previously we saw that a number is an eigenvalue of if and only if . This statement can be refined by means of the notion of minimal polynomial.
Let be a linear map from a finite-dimensional vector space to itself. A number is an eigenvalue of if and only if it is a root of the minimal polynomial of .
If is a real vector space, then, according to the Fundamental theorem of algebra, the characteristic polynomial of will factor into a product of linear polynomials and quadratic polynomials with leading coefficients and negative discriminants. Each of these quadratic polynomials will also occur as a divisor of the minimal polynomial (not necessarily with the same multiplicity).
If is a root of the minimal polynomial, then it also is a root of the characteristic polynomial (which, after all, is a multiple of the minimal polynomial), and so, according to rule 1 for the Characterization of eigenvalues and eigenvectors, an eigenvalue.
Conversely, if is an eigenvalue of with eigenvector , then, for the minimal polynomial , we have Since is an eigenvector, we have , so . This shows that is a root of the minimal polynomial.
It remains to prove the last statement. Suppose that is a real vector space and that is a quadratic polynomial with leading coefficient and negative discriminant which divides the characteristic polynomial of . If is a root of , then it is not real, and therefore the complex conjugate is the second root of . By applying the above to the complexification of , we see that both and occur as factors of the minimal polynomial of (the minimal polynomial of on the complexification of equals the minimal polynomial of on ). Because , the product is a divisor of the minimal polynomial.
The difference between the minimal polynomial and the characteristic polynomial is that factors occurring more than once in the characteristic polynomial may occur with a lower multiplicity in the minimal polynomial. The following three matrices illustrate this.
All three have characteristic polynomial . But
- has minimal polynomial ,
- has minimal polynomial ,
- has minimal polynomial .
These matrices are examples of matrices in Jordan normal form, which we will discuss later.
We use the minimal polynomial for the following characterization of diagonalizability.
Let be a vector space of finite dimension with basis and let be a linear map.
- If is a real vector space, then is diagonalizable (over the real numbers) if and only if the minimal polynomial of is a product of a constant and linear factors (with leading coefficients equal to ), which are all mutually different.
- If is a real vector space, then is diagonalizable over the complex numbers if and only if every factor of the minimal polynomial of (written with leading coefficient equal to ) occurs only once.
- If is a complex vector space, then is diagonalizable if and only if the minimal polynomial of has no double roots.
If is a diagonal matrix, and are the mutually different numbers on its diagonal, then the minimal polynomial of , and thus of , is equal to the product
Conversely, if, for mutually different numbers , the minimal polynomial of is equal to , then, because of the above theorem Roots of the minimal polynomial, these numbers are eigenvalues of . Furthermore, with , we have
This follows from the fact that the right-hand side is a polynomial of degree with the value at distinct points (see Lagrange's theorem).
We deduce that is the direct sum of the eigenspaces .
To this end, we first substitute into the formula above, and study the image of an arbitrary vector under the result:
This shows that is the sum of the linear subspaces .
We also see that
so is contained in .
If, for each , a basis of is chosen, then, according to theorem Independence of eigenvectors for different eigenvalues, the union of these bases is a set of linearly independent vectors. By what we saw above, these vectors span . Therefore, this union is a basis for . This shows that is the direct sum of the eigenspaces .
This proves the first and the last statement. The second statement follows by applying this result to the complexification of .
The matrix is not diagonalizable. For, otherwise there would be numbers and such that is conjugate to . But then
If we substitute from the first equation into the second equation, we find the quadratic equation , which has only one solution: . This implies , so , the identity matrix. That means that there is an invertible -matrix with . This contradicts the fact that the -entry of is equal to .
In accordance with the theorem, the minimal polynomial of has a double root.
- An orthogonal projection onto a subspace of an inner product space satisfies the equation and so is diagonalizable.
- A reflection about a linear subspace of dimension in an -dimensional inner product space satisfies the equation , and so is diagonalizable.
A well-known criterion for the absence of multiple factors in the factorization of a polynomial is Here stands for "greatest common divisor". With the aid of the Euclidean algorithm for polynomials this greatest common divisor can be found in an efficient manner. This leads to the following method for determining the diagonalizability of a square matrix :
- Determine the minimal polynomial of .
- Calculate .
- If this equals , then is diagonalizable over the complex numbers. If so and if is real, then is diagonalizable over the real numbers if and only if all of the roots of are real.
Let be a real vector space of dimension and suppose that is a linear map with minimal polynomial
Is diagonalizable over the real numbers?
Yes
To see this, we determine the eigenvalues of . Trial and error yields that is a root of . Division by gives the factorization
The quadratic factor factors as , so the minimal polynomial has three distinct real roots. According to the first item of theorem
Recognizing diagonalizability using the minimal polynomial the answer is therefore: Yes.