Undergraduate → Algebra → Linear Algebra ↓
Diagonalization
Diagonalization is a concept in linear algebra that involves transforming a square matrix into diagonal form. In simple terms, diagonalization is the process of finding a diagonal matrix that is similar to a given square matrix. This process is incredibly useful because diagonal matrices are the simplest form of matrix and are quite easy to work with, especially for calculations involving powers of matrices and solving differential equations.
Before getting into diagonalization, it is important to familiarize yourself with some basic concepts like matrix, eigenvectors, and eigenvalues, as they play a vital role in diagonalization.
Understanding matrices
A matrix is a rectangular array of numbers that can be used to represent linear transformations and systems of linear equations. A matrix with m
rows and n
columns is called mxn
matrix:
A = [a 11 a 12 a 13 a 21 a 22 a 23 a 31 a 32 a 33 ]
In this example, matrix A
is a 3 x 3
matrix.
Eigenvalues and eigenvectors
One of the key concepts of diagonalization is eigenvalue and eigenvector. Suppose A
is a square matrix. A non-zero vector v
is called an eigenvector of A
if there exists a scalar λ
(lambda) such that:
a * v = λ * v
In this equation, λ
is called the eigenvalue associated with the eigenvector v
. Essentially, when a matrix acts on one of its eigenvectors, it simply stretches or contracts it without changing its direction.
Diagonalization process
To determine if a matrix A
is diagonalizable, we need to see if it can be expressed as:
A = P * D * P -1
Here, matrix D
is a diagonal matrix and matrix P
is an invertible matrix whose columns are the eigenvectors of A
. The diagonal elements of D
are the eigenvalues of A
.
Let's outline the steps to diagonalize a matrix:
Step 1: Find the eigenvalues
The eigenvalues of the matrix A
are found by solving the characteristic polynomial:
det(A - λI) = 0
Here, I
is the identity matrix of the same size as A
, and λ
denotes the eigenvalues.
Step 2: Find the eigenvectors
Determine the eigenvectors by solving for each eigenvalue:
(A – λI)v = 0
The solutions v
are the eigenvectors associated with the eigenvalue λ
.
Step 3: Create matrices P
and D
Construct the matrix P
using the eigenvectors as columns. The diagonal matrix D
will have the eigenvalues on its diagonal.
Step 4: Verify diagonalization
Check that:
A = P * D * P -1
If this equation is true, then the matrix A
is diagonalizable.
Example
Consider the matrix:
A = [4 1 2 3]
Step 1: Find the eigenvalues
The characteristic polynomial is:
det(A - λI) = det([4-λ 1 2 3 - λ])
Calculate the determinant:
(4 – λ)(3 – λ) – (2)(1) = λ 2 – 7λ + 10
Set this equation to zero and solve for λ:
λ 2 - 7λ + 10 = 0
Let's factor this in:
(λ – 5)(λ – 2) = 0
Thus, λ = 5 and λ = 2 are the eigenvalues.
Step 2: Find the eigenvectors
For λ = 5, we solve:
(A – 5I)v = 0
[4-5 1 2 3-5]v = [-1 1 2 -2]
Solving the system, we find the eigenvectors:
v 1 = [1 1]
For λ = 2, we solve:
(A – 2I)v = 0
[4-2 1 2 3-2]v = [2 1 2 1]
Solving the system, we find the eigenvectors:
v 2 = [1 -2]
Step 3: Create matrices P
and D
Create the matrix P
and the diagonal matrix D
using the eigenvectors:
P = [1 1 1 -2]
D = [5 0 0 2]
Step 4: Verify diagonalization
Now verify that multiplying P
, D
, and P -1
gives the original matrix A
:
P * D * P -1 = [1 1 1 -2] * [5 0 0 2] * [1 1 1 -2]
This confirms that the matrix A
is diagonalizable.
Why diagonalization is useful
Diagonalization greatly simplifies many linear algebra calculations. For example, when a matrix is raised to a large power, diagonalization allows the process to be feasible and simple. Let's look at an example of raising our matrix A
to a power of 3:
Example: computational efficiency
Instead of multiplying A
by itself several times, A 3
is simplified using diagonalization:
A 3 = (P * D * P -1) 3 = P * D 3 * P -1
Calculate D 3
by raising each diagonal element of D
to the power of 3:
D 3 = [5 3 0 0 2 3] = [125 0 0 8]
Now calculate:
A 3 = P * D 3 * P -1
Computing this multiplication is generally straightforward. Diagonalization provides practical computational advantages in a variety of applications, such as transforming systems of linear equations, efficiently computing matrix exponents, simplifying linear transformations, and many more.
Limitations and conditions
While diagonalization greatly simplifies calculations, not every matrix can be diagonalized. Diagonalization is only possible if the matrix has enough linearly independent eigenvectors to form the matrix P
. Specifically, an nxn
matrix can only be diagonalized if it has n
linearly independent eigenvectors. Matrices that do not have these eigenvectors or have repeated eigenvalues, without corresponding eigenvectors that can form the required n
linearly independent sets, are not diagonalizable. These matrices are called defective matrices.
Understanding when diagonalization is possible is important in determining how to effectively simplify complex matrix operations. This process also extends beyond pure mathematics into applied fields such as physics, engineering, and computer science, where linear transformations are important.