In short

An eigenvalue of a square matrix A is a scalar \lambda such that A\mathbf{v} = \lambda\mathbf{v} for some nonzero vector \mathbf{v} (the eigenvector). The eigenvalues are the roots of the characteristic equation \det(A - \lambda I) = 0. The Cayley-Hamilton theorem says that every square matrix satisfies its own characteristic equation — if you plug the matrix itself into the characteristic polynomial, you get the zero matrix.

Take the matrix

A = \begin{bmatrix} 3 & 1 \\ 0 & 2 \end{bmatrix}

and multiply it by the vector \mathbf{v} = \begin{bmatrix} 1 \\ 0 \end{bmatrix}:

A\mathbf{v} = \begin{bmatrix} 3 & 1 \\ 0 & 2 \end{bmatrix} \begin{bmatrix} 1 \\ 0 \end{bmatrix} = \begin{bmatrix} 3 \\ 0 \end{bmatrix} = 3 \begin{bmatrix} 1 \\ 0 \end{bmatrix} = 3\mathbf{v}

The output is 3 times the input. The matrix A did not rotate \mathbf{v} or tilt it sideways — it just stretched it by a factor of 3, like pulling a rubber band along its own direction.

Now try \mathbf{w} = \begin{bmatrix} 1 \\ -1 \end{bmatrix}:

A\mathbf{w} = \begin{bmatrix} 3 & 1 \\ 0 & 2 \end{bmatrix} \begin{bmatrix} 1 \\ -1 \end{bmatrix} = \begin{bmatrix} 2 \\ -2 \end{bmatrix} = 2\begin{bmatrix} 1 \\ -1 \end{bmatrix} = 2\mathbf{w}

Again, the output is a scalar multiple of the input — this time by 2.

Most vectors get both rotated and stretched when you multiply them by a matrix. But these two — \mathbf{v} and \mathbf{w} — only get stretched (or compressed). They stay on the same line they started on. Vectors with this special property are called eigenvectors, and the scale factors (3 and 2) are called eigenvalues.

These numbers are not curiosities. The eigenvalues of a matrix encode the most fundamental information about what the matrix does — how much it stretches, whether it flips directions, whether it is stable or unstable in applications from differential equations to Google's PageRank algorithm to the vibration modes of a bridge. Finding them is one of the central problems in linear algebra.

Two eigenvectors of $A = \begin{bmatrix} 3 & 1 \\ 0 & 2 \end{bmatrix}$. The vector $\mathbf{v} = (1, 0)$ gets stretched to $3\mathbf{v}$ — same direction, three times the length (eigenvalue $3$). The vector $\mathbf{w} = (1, -1)$ gets stretched to $2\mathbf{w}$ (eigenvalue $2$). Both eigenvectors stay on their own line; the matrix only changes their magnitude, not their direction.

The equation A\mathbf{v} = \lambda\mathbf{v}

Eigenvalue and eigenvector

Let A be a square matrix of order n. A scalar \lambda is called an eigenvalue of A if there exists a nonzero vector \mathbf{v} such that

A\mathbf{v} = \lambda\mathbf{v}

The nonzero vector \mathbf{v} is called an eigenvector corresponding to the eigenvalue \lambda.

Two things to note.

The eigenvector must be nonzero. The zero vector trivially satisfies A\mathbf{0} = \lambda\mathbf{0} for every \lambda, so allowing it would make every number an eigenvalue, which is useless.

The eigenvalue can be zero. If \lambda = 0, the equation says A\mathbf{v} = \mathbf{0} — the matrix sends some nonzero vector to zero. This happens exactly when A is singular (non-invertible), which means \det A = 0. So zero is an eigenvalue of A if and only if A is not invertible.

Finding eigenvalues: the characteristic equation

How do you find the eigenvalues without guessing eigenvectors? Rearrange the defining equation:

A\mathbf{v} = \lambda\mathbf{v}
A\mathbf{v} - \lambda\mathbf{v} = \mathbf{0}
(A - \lambda I)\mathbf{v} = \mathbf{0}

This is a homogeneous system (A - \lambda I)\mathbf{v} = \mathbf{0}, and you need a nonzero solution \mathbf{v}. From the consistency of systems article, a homogeneous system has non-trivial solutions exactly when \text{rank}(A - \lambda I) < n, which for a square matrix means

\det(A - \lambda I) = 0

This is the key equation.

Characteristic equation

The characteristic equation of a square matrix A of order n is

\det(A - \lambda I) = 0

The left side, viewed as a polynomial in \lambda, is called the characteristic polynomial of A. For an n \times n matrix, the characteristic polynomial has degree n, so there are exactly n eigenvalues (counting multiplicity, and allowing complex values).

Computing the characteristic equation for a 2 \times 2 matrix

Take A = \begin{bmatrix} a & b \\ c & d \end{bmatrix}. Then

A - \lambda I = \begin{bmatrix} a - \lambda & b \\ c & d - \lambda \end{bmatrix}
\det(A - \lambda I) = (a - \lambda)(d - \lambda) - bc

Expanding:

= \lambda^2 - (a + d)\lambda + (ad - bc)
= \lambda^2 - \text{tr}(A)\,\lambda + \det(A)

where \text{tr}(A) = a + d is the trace (sum of diagonal entries) and \det(A) = ad - bc is the determinant. So for any 2 \times 2 matrix, the eigenvalues satisfy the quadratic

\lambda^2 - \text{tr}(A)\,\lambda + \det(A) = 0

This is a clean formula worth remembering. The sum of the eigenvalues equals the trace, and the product of the eigenvalues equals the determinant.

Computing the characteristic equation for a 3 \times 3 matrix

For A = \begin{bmatrix} a_{11} & a_{12} & a_{13} \\ a_{21} & a_{22} & a_{23} \\ a_{31} & a_{32} & a_{33} \end{bmatrix}, the characteristic polynomial is a cubic:

\det(A - \lambda I) = -\lambda^3 + \text{tr}(A)\,\lambda^2 - (\text{sum of cofactors of diagonal entries})\,\lambda + \det(A) = 0

In practice, you compute A - \lambda I explicitly and expand the determinant along a row or column, just as you would for any 3 \times 3 determinant. The result is a cubic equation in \lambda, which you solve by inspection, factoring, or the rational root theorem.

The characteristic polynomial $p(\lambda) = \lambda^2 - 7\lambda + 10$ for the matrix $A = \begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix}$ (used in Example 1 below). The eigenvalues are the roots — the values of $\lambda$ where the curve crosses the horizontal axis. Finding eigenvalues is equivalent to finding roots of a polynomial.

Worked examples

Example 1: Eigenvalues and eigenvectors of a 2×2 matrix

Find the eigenvalues and eigenvectors of A = \begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix}.

Step 1. Write A - \lambda I.

A - \lambda I = \begin{bmatrix} 4 - \lambda & 2 \\ 1 & 3 - \lambda \end{bmatrix}

Why: subtracting \lambda from each diagonal entry sets up the characteristic equation.

Step 2. Compute the determinant and set it to zero.

\det(A - \lambda I) = (4 - \lambda)(3 - \lambda) - 2 \cdot 1 = 12 - 7\lambda + \lambda^2 - 2 = \lambda^2 - 7\lambda + 10 = 0

Why: the determinant being zero is the condition for A\mathbf{v} = \lambda\mathbf{v} to have nonzero solutions.

Step 3. Solve the quadratic.

\lambda^2 - 7\lambda + 10 = (\lambda - 2)(\lambda - 5) = 0

So \lambda_1 = 2 and \lambda_2 = 5.

Why: the characteristic polynomial of a 2 \times 2 matrix is always a quadratic, so the eigenvalues are its roots. Check: \lambda_1 + \lambda_2 = 7 = \text{tr}(A) and \lambda_1 \cdot \lambda_2 = 10 = \det(A).

Step 4. Find eigenvectors.

For \lambda_1 = 2: solve (A - 2I)\mathbf{v} = \mathbf{0}.

A - 2I = \begin{bmatrix} 2 & 2 \\ 1 & 1 \end{bmatrix}

The two rows are proportional (row 1 is 2 \times row 2), so rank = 1. From row 2: v_1 + v_2 = 0, so v_2 = -v_1. Taking v_1 = 1:

\mathbf{v}_1 = \begin{bmatrix} 1 \\ -1 \end{bmatrix}

For \lambda_2 = 5: solve (A - 5I)\mathbf{v} = \mathbf{0}.

A - 5I = \begin{bmatrix} -1 & 2 \\ 1 & -2 \end{bmatrix}

Again rank = 1. From row 1: -v_1 + 2v_2 = 0, so v_1 = 2v_2. Taking v_2 = 1:

\mathbf{v}_2 = \begin{bmatrix} 2 \\ 1 \end{bmatrix}

Why: once you have the eigenvalue, the eigenvector comes from a homogeneous system. The rank drops to n - 1 (or lower), guaranteeing at least one free variable and hence a nonzero solution.

Result: Eigenvalues \lambda = 2 and \lambda = 5, with eigenvectors \begin{bmatrix} 1 \\ -1 \end{bmatrix} and \begin{bmatrix} 2 \\ 1 \end{bmatrix}.

The eigenvector $\mathbf{v}_1 = (1, -1)$ stretches by a factor of $2$ (red), and $\mathbf{v}_2 = (2, 1)$ stretches by a factor of $5$ (black). Both remain on their own line. Every other vector would be rotated as well as stretched — only these special directions are preserved.

The trace is 4 + 3 = 7, and the two eigenvalues sum to 2 + 5 = 7. The determinant is 4 \times 3 - 2 \times 1 = 10, and the two eigenvalues multiply to 2 \times 5 = 10. These are not coincidences — they hold for every 2 \times 2 matrix.

Example 2: Eigenvalues of a 3×3 matrix

Find the eigenvalues of A = \begin{bmatrix} 2 & 0 & 1 \\ 0 & 3 & 0 \\ 1 & 0 & 2 \end{bmatrix}.

Step 1. Write A - \lambda I.

A - \lambda I = \begin{bmatrix} 2-\lambda & 0 & 1 \\ 0 & 3-\lambda & 0 \\ 1 & 0 & 2-\lambda \end{bmatrix}

Why: same setup — subtract \lambda from the diagonal.

Step 2. Compute \det(A - \lambda I) by expanding along row 2, which has two zeros.

\det(A - \lambda I) = (3 - \lambda) \cdot \det\begin{bmatrix} 2 - \lambda & 1 \\ 1 & 2 - \lambda \end{bmatrix}
= (3 - \lambda)\bigl[(2 - \lambda)^2 - 1\bigr]
= (3 - \lambda)(4 - 4\lambda + \lambda^2 - 1)
= (3 - \lambda)(\lambda^2 - 4\lambda + 3)
= (3 - \lambda)(\lambda - 1)(\lambda - 3)

Why: expanding along a row with zeros minimises the computation. The inner quadratic factors cleanly.

Step 3. Set equal to zero.

(3 - \lambda)(\lambda - 1)(\lambda - 3) = 0

Note that (3 - \lambda) = -({\lambda - 3}), so this is -(\lambda - 3)^2(\lambda - 1) = 0.

The eigenvalues are \lambda = 1 and \lambda = 3 (with \lambda = 3 having algebraic multiplicity 2).

Step 4. Verify using trace and determinant.

Sum of eigenvalues: 1 + 3 + 3 = 7. Trace of A: 2 + 3 + 2 = 7. Matches.

Product of eigenvalues: 1 \times 3 \times 3 = 9. Determinant of A: expand along row 2 to get 3 \times (4 - 1) = 9. Matches.

Result: Eigenvalues are \lambda = 1 (simple) and \lambda = 3 (repeated, multiplicity 2).

Eigenvalue number line for the 3 by 3 matrixA number line showing the eigenvalues lambda equals 1 (marked with one dot) and lambda equals 3 (marked with a double dot to indicate multiplicity 2). The trace equals 7 and determinant equals 9 are labelled. 0 1 2 3 λ = 1 λ = 3 (×2) trace = 1 + 3 + 3 = 7 ✓ det = 1 × 3 × 3 = 9 ✓
The eigenvalues of $A$ on a number line. The double circle at $\lambda = 3$ indicates algebraic multiplicity $2$ — the factor $(\lambda - 3)$ appears squared in the characteristic polynomial. The trace and determinant serve as quick checks.

The repeated eigenvalue \lambda = 3 does not mean there is only one eigenvector — it means the characteristic polynomial has a double root. Whether there are one or two linearly independent eigenvectors for this eigenvalue depends on the rank of A - 3I. Here, A - 3I = \begin{bmatrix} -1 & 0 & 1 \\ 0 & 0 & 0 \\ 1 & 0 & -1 \end{bmatrix} has rank 1, so the null space is 2-dimensional: there are two independent eigenvectors for \lambda = 3.

Properties of eigenvalues

Several useful facts follow directly from the characteristic equation.

1. The sum of eigenvalues equals the trace. For any n \times n matrix, \lambda_1 + \lambda_2 + \cdots + \lambda_n = \text{tr}(A). You already verified this in both examples.

2. The product of eigenvalues equals the determinant. \lambda_1 \cdot \lambda_2 \cdots \lambda_n = \det(A). This is because the constant term of the characteristic polynomial is \det(A) (set \lambda = 0), and the product of the roots of a monic polynomial is (-1)^n times the constant term.

3. A matrix is singular if and only if zero is an eigenvalue. This is a direct consequence of property 2: \det A = 0 \iff one of the \lambda_i is zero.

4. The eigenvalues of A^{-1} are 1/\lambda_1, 1/\lambda_2, \ldots If A\mathbf{v} = \lambda\mathbf{v} and A is invertible, multiply both sides by A^{-1}: \mathbf{v} = \lambda A^{-1}\mathbf{v}, so A^{-1}\mathbf{v} = \frac{1}{\lambda}\mathbf{v}.

5. The eigenvalues of A^k are \lambda_1^k, \lambda_2^k, \ldots Apply A to A\mathbf{v} = \lambda\mathbf{v}: A^2\mathbf{v} = A(\lambda\mathbf{v}) = \lambda(A\mathbf{v}) = \lambda^2\mathbf{v}. Repeat for any power k.

Summary of eigenvalue propertiesA diagram showing how eigenvalue properties connect. A central box labelled A with eigenvalues lambda links to four derived matrices: A inverse has eigenvalues 1/lambda, A squared has eigenvalues lambda squared, A transpose has the same eigenvalues, and cA has eigenvalues c times lambda. A → λ eigenvalues A⁻¹ → 1/λ A² → λ² Aᵀ → λ cA → cλ
How eigenvalues transform under common matrix operations. If $A$ has eigenvalue $\lambda$, then $A^{-1}$ has eigenvalue $1/\lambda$, $A^2$ has eigenvalue $\lambda^2$, $A^T$ has the same eigenvalue $\lambda$, and $cA$ has eigenvalue $c\lambda$. These rules let you find eigenvalues of derived matrices without recomputing the characteristic equation.

The Cayley-Hamilton theorem

Here is a remarkable fact. Take the characteristic polynomial of a matrix — say, for A = \begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix}, the characteristic polynomial is p(\lambda) = \lambda^2 - 7\lambda + 10.

Now do something that looks illegal: replace \lambda with the matrix A itself:

p(A) = A^2 - 7A + 10I

The claim is that this always equals the zero matrix.

Compute it:

A^2 = \begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix}\begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix} = \begin{bmatrix} 18 & 14 \\ 7 & 11 \end{bmatrix}
7A = \begin{bmatrix} 28 & 14 \\ 7 & 21 \end{bmatrix}
10I = \begin{bmatrix} 10 & 0 \\ 0 & 10 \end{bmatrix}
A^2 - 7A + 10I = \begin{bmatrix} 18 - 28 + 10 & 14 - 14 + 0 \\ 7 - 7 + 0 & 11 - 21 + 10 \end{bmatrix} = \begin{bmatrix} 0 & 0 \\ 0 & 0 \end{bmatrix}

It works.

Cayley-Hamilton verification: A squared minus 7A plus 10I equals the zero matrixThree matrices labelled A squared, minus 7A, and plus 10I are shown side by side, with a final equals sign leading to the zero matrix. Each matrix shows its entries, and the sum of corresponding entries is zero. 18 14 7 11 7A 28 14 7 21 + 10I 10 0 0 10 = O 0 0 0 0 Every entry cancels to zero — the matrix satisfies its own characteristic equation.
The Cayley-Hamilton verification for $A = \begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix}$. The three matrices $A^2$, $7A$, and $10I$ combine entry-by-entry to produce the zero matrix. This is not a coincidence — it works for every square matrix.

Cayley-Hamilton theorem

Every square matrix satisfies its own characteristic equation. If p(\lambda) = \det(\lambda I - A) is the characteristic polynomial of A, then

p(A) = O

where O is the zero matrix.

For a 2 \times 2 matrix with characteristic polynomial \lambda^2 - \text{tr}(A)\,\lambda + \det(A) = 0, the theorem says

A^2 - \text{tr}(A) \cdot A + \det(A) \cdot I = O

For a 3 \times 3 matrix with characteristic polynomial \lambda^3 - c_2\lambda^2 + c_1\lambda - c_0 = 0, the theorem says

A^3 - c_2 A^2 + c_1 A - c_0 I = O

Verification for the 3 \times 3 example

Take A = \begin{bmatrix} 2 & 0 & 1 \\ 0 & 3 & 0 \\ 1 & 0 & 2 \end{bmatrix} from Example 2. The characteristic polynomial gives A^3 - 7A^2 + 15A - 9I = O.

Compute A^2 = \begin{bmatrix} 5 & 0 & 4 \\ 0 & 9 & 0 \\ 4 & 0 & 5 \end{bmatrix} and A^3 = \begin{bmatrix} 14 & 0 & 13 \\ 0 & 27 & 0 \\ 13 & 0 & 14 \end{bmatrix}.

Check entry (1,1): 14 - 35 + 30 - 9 = 0. Entry (1,3): 13 - 28 + 15 - 0 = 0. Entry (2,2): 27 - 63 + 45 - 9 = 0. Every entry vanishes — the 3 \times 3 matrix satisfies its own characteristic equation too.

Applications of the Cayley-Hamilton theorem

The Cayley-Hamilton theorem is not just a pretty identity — it has direct computational uses.

Finding A^{-1} from the characteristic equation

For a 2 \times 2 matrix, the Cayley-Hamilton theorem gives A^2 - \text{tr}(A) \cdot A + \det(A) \cdot I = O. If \det A \neq 0, rearrange:

A^2 - \text{tr}(A) \cdot A = -\det(A) \cdot I
A\bigl(A - \text{tr}(A) \cdot I\bigr) = -\det(A) \cdot I
A \cdot \frac{1}{-\det(A)}\bigl(A - \text{tr}(A) \cdot I\bigr) = I

So

A^{-1} = \frac{1}{\det(A)}\bigl(\text{tr}(A) \cdot I - A\bigr)

For A = \begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix}: \text{tr}(A) = 7, \det(A) = 10.

A^{-1} = \frac{1}{10}\left(\begin{bmatrix} 7 & 0 \\ 0 & 7 \end{bmatrix} - \begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix}\right) = \frac{1}{10}\begin{bmatrix} 3 & -2 \\ -1 & 4 \end{bmatrix}

Verify: A \cdot A^{-1} = \frac{1}{10}\begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix}\begin{bmatrix} 3 & -2 \\ -1 & 4 \end{bmatrix} = \frac{1}{10}\begin{bmatrix} 10 & 0 \\ 0 & 10 \end{bmatrix} = I. It works.

Computing higher powers of a matrix

The Cayley-Hamilton theorem lets you express A^n (for any n \geq 2) as a linear combination of I and A (for a 2 \times 2 matrix) or I, A, and A^2 (for 3 \times 3). The idea: use A^2 = \text{tr}(A) \cdot A - \det(A) \cdot I to replace A^2 wherever it appears. Then A^3 = A \cdot A^2 = A \cdot (\text{tr}(A) \cdot A - \det(A) \cdot I) = \text{tr}(A) \cdot A^2 - \det(A) \cdot A, and substitute again for A^2. This reduces any power to a linear combination of I and A.

For A = \begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix}: A^2 = 7A - 10I.

A^3 = A \cdot A^2 = A(7A - 10I) = 7A^2 - 10A = 7(7A - 10I) - 10A = 49A - 70I - 10A = 39A - 70I
A^3 = 39\begin{bmatrix} 4 & 2 \\ 1 & 3 \end{bmatrix} - 70\begin{bmatrix} 1 & 0 \\ 0 & 1 \end{bmatrix} = \begin{bmatrix} 156 - 70 & 78 \\ 39 & 117 - 70 \end{bmatrix} = \begin{bmatrix} 86 & 78 \\ 39 & 47 \end{bmatrix}

Without Cayley-Hamilton, computing A^3 requires two full matrix multiplications. With it, you only need scalar arithmetic once you have the relation A^2 = 7A - 10I.

Cayley-Hamilton power reduction chainA chain diagram showing how higher powers of a 2 by 2 matrix reduce. A squared equals 7A minus 10I. A cubed equals 39A minus 70I. A to the fourth equals 203A minus 490I. Each higher power reduces to a linear combination of A and I. A² = 7A−10I A³ = 39A−70I A⁴ = 203A−490I → ··· Every power of A is just (some number)·A + (some number)·I
The Cayley-Hamilton reduction chain. For a $2 \times 2$ matrix, every power $A^n$ reduces to a linear combination $\alpha A + \beta I$. The relation $A^2 = 7A - 10I$ is applied repeatedly — each step eliminates $A^2$ and higher, leaving only $A$ and $I$.

Common confusions

Going deeper

If you came here to learn what eigenvalues are, how to find them, and what the Cayley-Hamilton theorem says, you have it — you can stop here. The rest covers the proof idea behind Cayley-Hamilton, the connection between algebraic and geometric multiplicity, and why eigenvalues matter beyond pure mathematics.

Why Cayley-Hamilton is true (the idea)

The full proof of the Cayley-Hamilton theorem requires tools from linear algebra (such as the adjugate of A - \lambda I, whose entries are polynomials in \lambda) and is more involved than a single article can do justice to. But the intuition is accessible.

For each eigenvalue \lambda_i, the characteristic polynomial satisfies p(\lambda_i) = 0 by construction. If A has n linearly independent eigenvectors \mathbf{v}_1, \ldots, \mathbf{v}_n, then for each one:

p(A)\mathbf{v}_i = p(\lambda_i)\mathbf{v}_i = 0 \cdot \mathbf{v}_i = \mathbf{0}

This uses property 5 (eigenvalues of A^k are \lambda_i^k) and linearity. Since p(A) sends every basis vector to \mathbf{0}, it sends every vector to \mathbf{0} — meaning p(A) = O.

This argument works perfectly when A has n independent eigenvectors (the "diagonalisable" case). For matrices that are not diagonalisable, the proof requires more care — but the result still holds for every square matrix.

Algebraic vs. geometric multiplicity

The algebraic multiplicity of an eigenvalue is its multiplicity as a root of the characteristic polynomial. The geometric multiplicity is the dimension of its eigenspace — the number of linearly independent eigenvectors.

The geometric multiplicity is always at least 1 (every eigenvalue has at least one eigenvector) and at most the algebraic multiplicity. When the two are equal for every eigenvalue, the matrix is diagonalisable. When they differ for some eigenvalue, the matrix has a more complicated structure (described by its Jordan normal form).

In Example 2, \lambda = 3 has algebraic multiplicity 2. Its eigenspace is 2-dimensional (geometric multiplicity = 2), so the multiplicities match and the matrix is diagonalisable.

Why eigenvalues matter

Eigenvalues appear wherever matrices appear, and matrices appear everywhere.

Differential equations. The stability of a system of differential equations \mathbf{x}' = A\mathbf{x} is determined by the eigenvalues of A. If all eigenvalues have negative real part, all solutions decay to zero (the system is stable). If any eigenvalue has positive real part, solutions blow up.

Vibration analysis. The natural frequencies at which a structure vibrates are the square roots of the eigenvalues of a stiffness matrix. When an earthquake hits, the eigenvalues tell you which modes of oscillation the building will experience.

Principal Component Analysis. In data science and machine learning, PCA finds the directions of maximum variance in a dataset by computing the eigenvalues and eigenvectors of the covariance matrix. The largest eigenvalue tells you the direction in which the data varies most.

Where this leads next

Eigenvalues are a gateway to the deeper structure of linear algebra. The most direct continuations: