Skip to main content

Eigenvalues and Eigenvectors

Eigenvectors reveal directions that a matrix does not turn. Along those directions, the matrix only stretches, compresses, or reverses orientation. Eigenvalues measure that scalar action. This idea explains long-term dynamics, diagonalization, powers of matrices, differential equations, Markov chains, and spectral geometry.

The definition is simple, but its consequences are broad. Instead of asking how a matrix acts on every vector at once, eigenanalysis looks for special directions where the action is one-dimensional. If enough such directions exist, the matrix can be understood by studying independent scalar multiplications.

Definitions

Let AA be an n×nn\times n matrix. A nonzero vector x\mathbf{x} is an eigenvector of AA if

Ax=λxA\mathbf{x}=\lambda\mathbf{x}

for some scalar λ\lambda. The scalar λ\lambda is the corresponding eigenvalue.

Equivalently,

(AλI)x=0.(A-\lambda I)\mathbf{x}=\mathbf{0}.

Thus λ\lambda is an eigenvalue exactly when AλIA-\lambda I is singular:

det(AλI)=0.\det(A-\lambda I)=0.

The polynomial

pA(λ)=det(λIA)p_A(\lambda)=\det(\lambda I-A)

or equivalently det(AλI)\det(A-\lambda I) up to sign, is the characteristic polynomial. The eigenspace for λ\lambda is

Eλ=null(AλI).E_\lambda=\operatorname{null}(A-\lambda I).

Because eigenvectors are required to be nonzero, the zero vector is not itself an eigenvector. However, the eigenspace includes the zero vector because it is a subspace.

Key results

The eigenvalues of AA are the roots of the characteristic polynomial. For an n×nn\times n matrix, the characteristic polynomial has degree nn, though it may have repeated roots or complex roots.

If AA is triangular, its eigenvalues are the diagonal entries. This follows because AλIA-\lambda I is triangular, and its determinant is the product of diagonal entries:

det(AλI)=(a11λ)(a22λ)(annλ).\det(A-\lambda I)= (a_{11}-\lambda)(a_{22}-\lambda)\cdots(a_{nn}-\lambda).

Eigenvectors corresponding to distinct eigenvalues are linearly independent. A proof sketch for two eigenvectors is direct. Suppose Av1=λ1v1A\mathbf{v}_1=\lambda_1\mathbf{v}_1 and Av2=λ2v2A\mathbf{v}_2=\lambda_2\mathbf{v}_2, with λ1λ2\lambda_1\neq\lambda_2. If c1v1+c2v2=0c_1\mathbf{v}_1+c_2\mathbf{v}_2=\mathbf{0}, apply AA and compare with multiplying the original equation by λ2\lambda_2:

c1λ1v1+c2λ2v2=0,c_1\lambda_1\mathbf{v}_1+c_2\lambda_2\mathbf{v}_2=\mathbf{0},

and

c1λ2v1+c2λ2v2=0.c_1\lambda_2\mathbf{v}_1+c_2\lambda_2\mathbf{v}_2=\mathbf{0}.

Subtracting gives c1(λ1λ2)v1=0c_1(\lambda_1-\lambda_2)\mathbf{v}_1=\mathbf{0}, so c1=0c_1=0, and then c2=0c_2=0. The general case is similar.

The determinant and trace encode eigenvalue information. For a 2×22\times2 matrix, the characteristic polynomial can be written

λ2tr(A)λ+det(A),\lambda^2-\operatorname{tr}(A)\lambda+\det(A),

so the sum of eigenvalues is the trace and the product is the determinant, counting algebraic multiplicity.

The eigenspace EλE_\lambda is a subspace because it is the null space of AλIA-\lambda I. This matters when solving by hand: after finding an eigenvalue, one should row-reduce AλIA-\lambda I and describe the full solution space, not just one vector. Any nonzero vector in that subspace is an eigenvector, and any scalar multiple represents the same eigendirection.

Eigenvalues can be zero. A zero eigenvalue means there is a nonzero vector x\mathbf{x} such that Ax=0A\mathbf{x}=\mathbf{0}. Therefore AA has a nontrivial null space and is singular. Conversely, if AA is singular, then det(A)=0\det(A)=0, so λ=0\lambda=0 is a root of det(AλI)\det(A-\lambda I) up to sign. This gives a useful bridge between invertibility and spectral information.

Complex eigenvalues also have real consequences. A real 2×22\times2 rotation matrix by an angle that is not 00 or π\pi has no real eigenvectors because no real direction is left on its own line. Over the complex numbers it has complex eigenvalues. In applications, complex eigenvalues often indicate rotation or oscillation, while their magnitudes indicate growth or decay.

For repeated eigenvalues, the characteristic polynomial alone does not tell the whole story. If λ\lambda has algebraic multiplicity 33, its eigenspace might have dimension 11, 22, or 33. The dimension of the eigenspace determines how many independent eigenvectors are available for diagonalization. Thus the workflow is always: find eigenvalues, then find eigenspaces, then count independent eigenvectors.

In dynamical systems, the absolute values of eigenvalues control long-term behavior when the matrix is diagonalizable or nearly so. Eigenvalues with λ<1\vert \lambda\vert \lt 1 correspond to decaying modes. Eigenvalues with λ>1\vert \lambda\vert \gt 1 correspond to growing modes. Eigenvalues with negative sign introduce alternating direction. This modal interpretation is one of the main reasons eigenvectors are more than a calculation exercise.

Visual

ObjectHow it is foundMeaning
Eigenvalue λ\lambdaroot of det(AλI)=0\det(A-\lambda I)=0scalar stretch factor
Eigenvector x\mathbf{x}nonzero solution of (AλI)x=0(A-\lambda I)\mathbf{x}=0direction preserved by AA
Eigenspace EλE_\lambdanull space of AλIA-\lambda Iall vectors stretched by λ\lambda plus zero
Algebraic multiplicitymultiplicity as polynomial rootcharacteristic-polynomial count
Geometric multiplicitydim(Eλ)\dim(E_\lambda)number of independent eigenvectors for λ\lambda

Worked example 1: Find eigenvalues and eigenvectors of a 2 by 2 matrix

Problem: find the eigenvalues and eigenspaces of

A=[4123].A= \begin{bmatrix} 4&1\\ 2&3 \end{bmatrix}.

Step 1: compute the characteristic equation.

AλI=[4λ123λ].A-\lambda I= \begin{bmatrix} 4-\lambda&1\\ 2&3-\lambda \end{bmatrix}.

Thus

det(AλI)=(4λ)(3λ)2=127λ+λ22=λ27λ+10.\det(A-\lambda I) =(4-\lambda)(3-\lambda)-2 =12-7\lambda+\lambda^2-2 =\lambda^2-7\lambda+10.

Step 2: factor.

λ27λ+10=(λ5)(λ2).\lambda^2-7\lambda+10=(\lambda-5)(\lambda-2).

So the eigenvalues are λ=5\lambda=5 and λ=2\lambda=2.

Step 3: find E5E_5. Solve (A5I)x=0(A-5I)\mathbf{x}=\mathbf{0}:

[1122][xy]=[00].\begin{bmatrix} -1&1\\ 2&-2 \end{bmatrix} \begin{bmatrix}x\\y\end{bmatrix} = \begin{bmatrix}0\\0\end{bmatrix}.

The equation is x+y=0-x+y=0, so y=xy=x. Therefore

E5=span{[11]}.E_5=\operatorname{span}\left\{ \begin{bmatrix}1\\1\end{bmatrix} \right\}.

Step 4: find E2E_2. Solve (A2I)x=0(A-2I)\mathbf{x}=\mathbf{0}:

[2121][xy]=[00].\begin{bmatrix} 2&1\\ 2&1 \end{bmatrix} \begin{bmatrix}x\\y\end{bmatrix} = \begin{bmatrix}0\\0\end{bmatrix}.

The equation is 2x+y=02x+y=0, so y=2xy=-2x. Therefore

E2=span{[12]}.E_2=\operatorname{span}\left\{ \begin{bmatrix}1\\-2\end{bmatrix} \right\}.

Checked answer: A[11]=[55]A\begin{bmatrix}1\\1\end{bmatrix}=\begin{bmatrix}5\\5\end{bmatrix} and A[12]=[24]A\begin{bmatrix}1\\-2\end{bmatrix}=\begin{bmatrix}2\\-4\end{bmatrix}.

Worked example 2: Interpret eigenvectors in a repeated process

Problem: suppose a process updates states by

xk+1=Axk,A=[0.80.30.20.7].\mathbf{x}_{k+1}=A\mathbf{x}_k, \qquad A= \begin{bmatrix} 0.8&0.3\\ 0.2&0.7 \end{bmatrix}.

Find a steady direction, meaning an eigenvector for λ=1\lambda=1.

Step 1: solve (AI)x=0(A-I)\mathbf{x}=\mathbf{0}.

AI=[0.20.30.20.3].A-I= \begin{bmatrix} -0.2&0.3\\ 0.2&-0.3 \end{bmatrix}.

The equation is

0.2x+0.3y=0.-0.2x+0.3y=0.

Step 2: clear decimals:

2x+3y=03y=2xy=23x.-2x+3y=0 \quad\Longrightarrow\quad 3y=2x \quad\Longrightarrow\quad y=\frac23x.

Step 3: choose x=3x=3, giving y=2y=2. A steady eigenvector is

v=[32].\mathbf{v}= \begin{bmatrix} 3\\2 \end{bmatrix}.

Step 4: normalize to a probability vector if desired:

13+2[32]=[3/52/5].\frac{1}{3+2} \begin{bmatrix} 3\\2 \end{bmatrix} = \begin{bmatrix} 3/5\\2/5 \end{bmatrix}.

Checked answer:

A[32]=[0.8(3)+0.3(2)0.2(3)+0.7(2)]=[32].A\begin{bmatrix}3\\2\end{bmatrix} = \begin{bmatrix} 0.8(3)+0.3(2)\\ 0.2(3)+0.7(2) \end{bmatrix} = \begin{bmatrix} 3\\2 \end{bmatrix}.

The direction is unchanged by the update.

Code

import numpy as np

A = np.array([[4, 1],
[2, 3]], dtype=float)

values, vectors = np.linalg.eig(A)
print(values)
print(vectors)

for i, lam in enumerate(values):
v = vectors[:, i]
print(np.allclose(A @ v, lam * v))

Numerical eigenvectors are usually scaled differently from hand-computed eigenvectors. Any nonzero scalar multiple of an eigenvector is still an eigenvector for the same eigenvalue.

Common pitfalls

  • Allowing the zero vector as an eigenvector. It is in every eigenspace but is not called an eigenvector.
  • Solving det(AλI)=0\det(A-\lambda I)=0 and stopping before finding eigenspaces.
  • Confusing algebraic multiplicity with geometric multiplicity.
  • Assuming every real matrix has real eigenvalues. Some real matrices have complex eigenvalues.
  • Forgetting that eigenvectors are directions: scalar multiples represent the same eigendirection.
  • Using det(λA)\det(\lambda-A) instead of det(λIA)\det(\lambda I-A). The identity matrix is required.

A reliable eigenvalue workflow is sequential. First compute the characteristic polynomial carefully. Second solve for candidate eigenvalues. Third, for each eigenvalue, row-reduce AλIA-\lambda I and describe the null space. Fourth, check at least one eigenvector by multiplying AvA\mathbf{v} and comparing it with λv\lambda\mathbf{v}. Skipping the final check is risky because sign errors in characteristic polynomials are common.

When a problem asks for "the eigenvectors" for an eigenvalue, it usually expects the eigenspace, not just one vector. Since every nonzero scalar multiple is also an eigenvector, listing a single vector without saying "span" hides the full answer. A clean response is

Eλ=span{v1,,vk}.E_\lambda=\operatorname{span}\{\mathbf{v}_1,\ldots,\mathbf{v}_k\}.

Then the eigenvectors are the nonzero vectors in that span.

Repeated eigenvalues require extra care. A repeated root of the characteristic polynomial may produce one independent eigenvector or several. The only way to know is to solve the null-space problem. This distinction controls diagonalization, powers of matrices, and qualitative dynamics.

In applied settings, eigenvectors often represent modes. A mode is a pattern that keeps its shape while its size changes. In a population model, a steady distribution is an eigenvector for eigenvalue 11. In a vibration model, eigenvectors describe natural shapes of motion. In a data covariance matrix, eigenvectors identify principal directions of variation. The same algebraic equation supports all of these interpretations.

Eigenvalue checks can use determinant and trace for small matrices. For a 2×22\times2 matrix, if you find eigenvalues λ1\lambda_1 and λ2\lambda_2, then their sum should equal the trace and their product should equal the determinant. These checks do not replace the computation, but they are excellent at catching arithmetic errors.

For triangular matrices, do not expand the characteristic determinant unnecessarily. The eigenvalues are already on the diagonal because AλIA-\lambda I remains triangular. This includes diagonal matrices as the simplest case. The eigenspaces still require solving (AλI)x=0(A-\lambda I)\mathbf{x}=\mathbf{0}; only the eigenvalue step is immediate.

If a matrix is symmetric, its eigenvectors have extra structure: eigenvectors from distinct eigenspaces are orthogonal, and an orthonormal eigenbasis exists. That stronger result belongs to the spectral theorem, but it is useful to remember because symmetric eigenvalue problems are common and especially well behaved.

Connections