Английская Википедия:Characteristic polynomial

Материал из Онлайн справочника
Перейти к навигацииПерейти к поиску

Шаблон:Use American English Шаблон:Short description Шаблон:About

In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The characteristic polynomial of an endomorphism of a finite-dimensional vector space is the characteristic polynomial of the matrix of that endomorphism over any base (that is, the characteristic polynomial does not depend on the choice of a basis). The characteristic equation, also known as the determinantal equation,[1][2][3] is the equation obtained by equating the characteristic polynomial to zero.

In spectral graph theory, the characteristic polynomial of a graph is the characteristic polynomial of its adjacency matrix.[4]

Motivation

In linear algebra, eigenvalues and eigenvectors play a fundamental role, since, given a linear transformation, an eigenvector is a vector whose direction is not changed by the transformation, and the corresponding eigenvalue is the measure of the resulting change of magnitude of the vector.

More precisely, if the transformation is represented by a square matrix <math>A,</math> an eigenvector <math>\mathbf{v},</math> and the corresponding eigenvalue <math>\lambda</math> must satisfy the equation <math display=block>A \mathbf{v} = \lambda \mathbf{v},</math> or, equivalently, <math display=block>(\lambda I - A) \mathbf{v} =\mathbf 0</math> where <math>I</math> is the identity matrix, and <math>\mathbf{v}\ne \mathbf{0}</math> (although the zero vector satisfies this equation for every <math>\lambda,</math> it is not considered an eigenvector).

It follows that the matrix <math>(\lambda I - A)</math> must be singular, and its determinant <math display=block>\det(\lambda I - A) = 0</math> must be zero.

In other words, the eigenvalues of Шаблон:Mvar are the roots of <math display=block>\det(xI - A),</math> which is a monic polynomial in Шаблон:Mvar of degree Шаблон:Mvar if Шаблон:Mvar is a Шаблон:Math matrix. This polynomial is the characteristic polynomial of Шаблон:Mvar.

Formal definition

Consider an <math>n \times n</math> matrix <math>A.</math> The characteristic polynomial of <math>A,</math> denoted by <math>p_A(t),</math> is the polynomial defined by[5] <math display=block>p_A(t) = \det (t I - A)</math> where <math>I</math> denotes the <math>n \times n</math> identity matrix.

Some authors define the characteristic polynomial to be <math>\det(A - t I).</math> That polynomial differs from the one defined here by a sign <math>(-1)^n,</math> so it makes no difference for properties like having as roots the eigenvalues of <math>A</math>; however the definition above always gives a monic polynomial, whereas the alternative definition is monic only when <math>n</math> is even.

Examples

To compute the characteristic polynomial of the matrix <math display=block>A = \begin{pmatrix} 2 & 1\\ -1& 0 \end{pmatrix}. </math> the determinant of the following is computed: <math display=block>t I-A = \begin{pmatrix} t-2&-1\\ 1&t-0 \end{pmatrix} </math> and found to be <math>(t-2)t - 1(-1) = t^2-2t+1 \,\!,</math> the characteristic polynomial of <math>A.</math>

Another example uses hyperbolic functions of a hyperbolic angle φ. For the matrix take <math display=block>A = \begin{pmatrix} \cosh(\varphi) & \sinh(\varphi)\\ \sinh(\varphi)& \cosh(\varphi) \end{pmatrix}.</math> Its characteristic polynomial is <math display=block>\det (tI - A) = (t - \cosh(\varphi))^2 - \sinh^2(\varphi) = t^2 - 2 t \ \cosh(\varphi) + 1 = (t - e^\varphi) (t - e^{-\varphi}).</math>

Properties

The characteristic polynomial <math>p_A(t)</math> of a <math>n \times n</math> matrix is monic (its leading coefficient is <math>1</math>) and its degree is <math>n.</math> The most important fact about the characteristic polynomial was already mentioned in the motivational paragraph: the eigenvalues of <math>A</math> are precisely the roots of <math>p_A(t)</math> (this also holds for the minimal polynomial of <math>A,</math> but its degree may be less than <math>n</math>). All coefficients of the characteristic polynomial are polynomial expressions in the entries of the matrix. In particular its constant coefficient of <math>t^0</math> is <math>\det(-A) = (-1)^n \det(A),</math> the coefficient of <math>t^n</math> is one, and the coefficient of <math>t^{n-1}</math> is Шаблон:Math, where Шаблон:Math is the trace of <math>A.</math> (The signs given here correspond to the formal definition given in the previous section;[6] for the alternative definition these would instead be <math>\det(A)</math> and Шаблон:Math respectively.[7])

For a <math>2 \times 2</math> matrix <math>A,</math> the characteristic polynomial is thus given by <math display=block>t^2 - \operatorname{tr}(A) t + \det(A).</math>

Using the language of exterior algebra, the characteristic polynomial of an <math>n \times n</math> matrix <math>A</math> may be expressed as <math display=block>p_A (t) = \sum_{k=0}^n t^{n-k} (-1)^k \operatorname{tr}\left(\textstyle\bigwedge^k A\right)</math> where <math display="inline">\operatorname{tr}\left(\bigwedge^k A\right)</math> is the trace of the <math>k</math>th exterior power of <math>A,</math> which has dimension <math display="inline">\binom {n}{k}.</math> This trace may be computed as the sum of all principal minors of <math>A</math> of size <math>k.</math> The recursive Faddeev–LeVerrier algorithm computes these coefficients more efficiently.

When the characteristic of the field of the coefficients is <math>0,</math> each such trace may alternatively be computed as a single determinant, that of the <math>k \times k</math> matrix, <math display=block>\operatorname{tr}\left(\textstyle\bigwedge^k A\right) = \frac{1}{k!} \begin{vmatrix} \operatorname{tr}A & k-1 &0&\cdots &0 \\ \operatorname{tr}A^2 &\operatorname{tr}A& k-2 &\cdots &0 \\

\vdots & \vdots & & \ddots & \vdots    \\

\operatorname{tr}A^{k-1} &\operatorname{tr}A^{k-2}& & \cdots & 1 \\ \operatorname{tr}A^k &\operatorname{tr}A^{k-1}& & \cdots & \operatorname{tr}A \end{vmatrix} ~.</math>

The Cayley–Hamilton theorem states that replacing <math>t</math> by <math>A</math> in the characteristic polynomial (interpreting the resulting powers as matrix powers, and the constant term <math>c</math> as <math>c</math> times the identity matrix) yields the zero matrix. Informally speaking, every matrix satisfies its own characteristic equation. This statement is equivalent to saying that the minimal polynomial of <math>A</math> divides the characteristic polynomial of <math>A.</math>

Two similar matrices have the same characteristic polynomial. The converse however is not true in general: two matrices with the same characteristic polynomial need not be similar.

The matrix <math>A</math> and its transpose have the same characteristic polynomial. <math>A</math> is similar to a triangular matrix if and only if its characteristic polynomial can be completely factored into linear factors over <math>K</math> (the same is true with the minimal polynomial instead of the characteristic polynomial). In this case <math>A</math> is similar to a matrix in Jordan normal form.

Characteristic polynomial of a product of two matrices

If <math>A</math> and <math>B</math> are two square <math>n \times n</math> matrices then characteristic polynomials of <math>AB</math> and <math>BA</math> coincide: <math display=block>p_{AB}(t)=p_{BA}(t).\,</math>

When <math>A</math> is non-singular this result follows from the fact that <math>AB</math> and <math>BA</math> are similar: <math display=block>BA = A^{-1} (AB) A.</math>

For the case where both <math>A</math> and <math>B</math> are singular, the desired identity is an equality between polynomials in <math>t</math> and the coefficients of the matrices. Thus, to prove this equality, it suffices to prove that it is verified on a non-empty open subset (for the usual topology, or, more generally, for the Zariski topology) of the space of all the coefficients. As the non-singular matrices form such an open subset of the space of all matrices, this proves the result.

More generally, if <math>A</math> is a matrix of order <math>m \times n</math> and <math>B</math> is a matrix of order <math>n \times m,</math> then <math>AB</math> is <math>m \times m</math> and <math>BA</math> is <math>n \times n</math> matrix, and one has <math display=block>p_{BA}(t) = t^{n-m} p_{AB}(t).\,</math>

To prove this, one may suppose <math>n > m,</math> by exchanging, if needed, <math>A</math> and <math>B.</math> Then, by bordering <math>A</math> on the bottom by <math>n - m</math> rows of zeros, and <math>B</math> on the right, by, <math>n - m</math> columns of zeros, one gets two <math>n \times n</math> matrices <math>A^{\prime}</math> and <math>B^{\prime}</math> such that <math>B^{\prime}A^{\prime} = BA</math> and <math>A^{\prime}B^{\prime}</math> is equal to <math>AB</math> bordered by <math>n - m</math> rows and columns of zeros. The result follows from the case of square matrices, by comparing the characteristic polynomials of <math>A^{\prime}B^{\prime}</math> and <math>AB.</math>

Characteristic polynomial of Ak

If <math>\lambda</math> is an eigenvalue of a square matrix <math>A</math> with eigenvector <math>\mathbf{v},</math> then <math>\lambda^k</math> is an eigenvalue of <math>A^k</math> because <math display=block>A^k \textbf{v} = A^{k-1} A \textbf{v} = \lambda A^{k-1} \textbf{v} = \dots = \lambda^k \textbf{v}.</math>

The multiplicities can be shown to agree as well, and this generalizes to any polynomial in place of <math>x^k</math>:[8] Шаблон:Math theorem That is, the algebraic multiplicity of <math>\lambda</math> in <math>f(A)</math> equals the sum of algebraic multiplicities of <math>\lambda'</math> in <math>A</math> over <math>\lambda'</math> such that <math>f(\lambda') = \lambda.</math> In particular, <math>\operatorname{tr}(f(A)) = \textstyle\sum_{i=1}^n f(\lambda_i)</math> and <math>\operatorname{det}(f(A)) = \textstyle\prod_{i=1}^n f(\lambda_i).</math> Here a polynomial <math>f(t) = t^3+1,</math> for example, is evaluated on a matrix <math>A</math> simply as <math>f(A) = A^3+I.</math>

The theorem applies to matrices and polynomials over any field or commutative ring.[9] However, the assumption that <math>p_A(t)</math> has a factorization into linear factors is not always true, unless the matrix is over an algebraically closed field such as the complex numbers.

Шаблон:Math proof

Secular function and secular equation

Secular function

The term secular function has been used for what is now called characteristic polynomial (in some literature the term secular function is still used). The term comes from the fact that the characteristic polynomial was used to calculate secular perturbations (on a time scale of a century, that is, slow compared to annual motion) of planetary orbits, according to Lagrange's theory of oscillations.

Secular equation

Secular equation may have several meanings.

  • In linear algebra it is sometimes used in place of characteristic equation.
  • In astronomy it is the algebraic or numerical expression of the magnitude of the inequalities in a planet's motion that remain after the inequalities of a short period have been allowed for.[10]
  • In molecular orbital calculations relating to the energy of the electron and its wave function it is also used instead of the characteristic equation.

For general associative algebras

The above definition of the characteristic polynomial of a matrix <math>A \in M_n(F)</math> with entries in a field <math>F</math> generalizes without any changes to the case when <math>F</math> is just a commutative ring. Шаблон:Harvtxt defines the characteristic polynomial for elements of an arbitrary finite-dimensional (associative, but not necessarily commutative) algebra over a field <math>F</math> and proves the standard properties of the characteristic polynomial in this generality.

See also

References

Шаблон:Reflist