Today, I'm going to talk about complex eigenvalues - how we deal with them algebraically, how we interpret them geometrically, and how they increase the scope and power of the diagonalization techniques that we learned in the last presentation.
This stuff is a bit strewn out through the textbooks, as it includes examples with complex eigenvalues as it goes along. Here are a few examples of where you can find info in the text:
We will use some of the very basics of complex numbers. We'll need to know that a complex number has the form $a+bi$, where $a$ and $b$ are real numbers and $i$ denotes $\sqrt{-1}$.
We'll also need to know how complex numbers as solutions of quadratics.
If we want to solve $x^2+x+1=0$, we can apply the quadratic formula to get
$$\begin{align} x &= \frac{-1\pm\sqrt{1-4}}{2} \\ &= -\frac{1}{2} \pm \frac{\sqrt{-3}}{2} \\ &= -\frac{1}{2} \pm \frac{\sqrt{3}}{2}i. \end{align}$$Let $A = \begin{pmatrix}0 & -1 \\ 1 & 0\end{pmatrix}$, then
$$ \det(A-\lambda I) = \left|\begin{matrix} -\lambda & -1 \\ 1 & -\lambda \end{matrix} \right| = \lambda^2+1. $$The roots of the characteristic polynomial $\lambda^2+1$ are exactly $\pm i$.
The two standard coordinate basis vectors rotate through the angle $90^{\circ}$ under multiplication by $A$ since
$$\begin{align} \begin{pmatrix}0 & -1 \\ 1 & 0\end{pmatrix} \begin{pmatrix}1 \\ 0\end{pmatrix} &= \begin{pmatrix}0 \\ 1\end{pmatrix} \: \text{ and } \\ \begin{pmatrix}0 & -1 \\ 1 & 0\end{pmatrix} \begin{pmatrix}0 \\ 1\end{pmatrix} &= \begin{pmatrix}-1 \\ 0\end{pmatrix}. \end{align}$$This can be extended to any vector since
$$\begin{align} \begin{pmatrix}0 & -1 \\ 1 & 0\end{pmatrix} \begin{pmatrix}a \\ b\end{pmatrix} &= \begin{pmatrix}0 & -1 \\ 1 & 0\end{pmatrix} \left(a \begin{pmatrix}1 \\ 0\end{pmatrix} + b \begin{pmatrix}0 \\ 1\end{pmatrix}\right) \\ &= a \begin{pmatrix}0 & -1 \\ 1 & 0\end{pmatrix} \begin{pmatrix}1 \\ 0\end{pmatrix} + b \begin{pmatrix}0 & -1 \\ 1 & 0\end{pmatrix} \begin{pmatrix}0 \\ 1\end{pmatrix} \\ &= a \begin{pmatrix}0 \\ 1\end{pmatrix} + b \begin{pmatrix}-1 \\ 0\end{pmatrix} = \begin{pmatrix}-b \\ a\end{pmatrix} \end{align}$$Putting this all together, multiplication by $A=\begin{pmatrix}0 & -1 \\ 1 & 0\end{pmatrix}$ induces rotation through $90^{\circ}$.
Rotation through any angle $\theta$ can be represented via multiplication by
$$R_{\theta} = \begin{pmatrix} \cos(\theta)&-\sin(\theta) \\ \sin(\theta)&\cos(\theta) \end{pmatrix}.$$The reason is the same:
$$\begin{align} \begin{pmatrix} \cos(\theta)&-\sin(\theta) \\ \sin(\theta)&\cos(\theta) \end{pmatrix} \begin{pmatrix}1 \\ 0\end{pmatrix} &= \begin{pmatrix}\cos(\theta) \\ \sin(\theta)\end{pmatrix} \: \text{ and } \\ \begin{pmatrix} \cos(\theta)&-\sin(\theta) \\ \sin(\theta)&\cos(\theta) \end{pmatrix} \begin{pmatrix}0 \\ 1\end{pmatrix} &= \begin{pmatrix}-\sin(\theta) \\ \cos(\theta)\end{pmatrix}. \end{align}$$Thus, multiplication by $A_{\theta}$ rotates $\vec{\imath}$ and $\vec{\jmath}$ through the angle $\theta$ and (by linear extension) any vector.
Let's check the eigenvalues of $A_{\theta}$:
$$\begin{align} \det(A_{\theta}-\lambda I) &= \left|\begin{matrix} \cos(\theta)-\lambda & -\sin(\theta) \\ \sin(\theta) & \cos(\theta)-\lambda \end{matrix} \right| \\ &= \cos^2(\theta) - 2\cos(\theta)\lambda + \lambda^2 + \sin^2(\theta) \\ &= \lambda^2 - 2\cos(\theta)\lambda + 1. \end{align}$$We can compute the roots of $\lambda^2 - 2\cos(\theta)\lambda + 1$ by applying the quadratic formula. We get:
$$\begin{align} \lambda &= \frac{2\cos(\theta)\pm\sqrt{4\cos^2(\theta)-4}}{2} \\ &= \cos(\theta) \pm i \sin(\theta). \end{align}$$Complex eigenvalues indicate rotation!
Given a line through the origin and a point $P$ not on the line, how can we describe the reflection $P'$ of that point?
First note, that reflection across the $x$-axis is easy; it's just represented by
$$T = \begin{pmatrix}1&0\\0&-1\end{pmatrix}.$$Thus, reflection across the line through the origin making the angle $\theta$ with the $x$-axis is represented by
$$M = R_{\theta}TR_{-\theta}.$$We again start with an example:
$$M = \begin{pmatrix} 0 & 0 & 1 \\ 1 & 0 & 0 \\ 0 & 1 & 0 \end{pmatrix}.$$Note that $M$ maps $\vec{\imath}\to\vec{\jmath}\to\vec{k}\to\vec{\imath}$.
Since $M$ maps $\vec{\imath}\to\vec{\jmath}\to\vec{k}\to\vec{\imath}$, it's easy to imagine that it might represent rotation.
So, what's the eigenstructure? If we compute the characteristic polynomial, we get:
$$\left|\begin{matrix} -\lambda & 0 & 1 \\ 1 & -\lambda & 0 \\ 0 & 1 & -\lambda \end{matrix}\right| = 1-\lambda^3.$$Clearly, $\lambda=1$ is a root. You can solve $M\vec{v}=\vec{v}$ to show that $\langle 1,1,1 \rangle$ is a corresponding eigenvector. This represents the fixed direction around which the rotation happens.
To find the other two roots, we factor:
$$1-\lambda^3 = (1-\lambda ) \left(\lambda ^2+\lambda +1\right).$$The roots of that second factor are
$$\begin{align}\lambda &= \frac{-1\pm\sqrt{1-4}}{2} = -\frac{1}{2}\pm\frac{\sqrt{3}}{2}i \\ &= \cos\left(\frac{2\pi}{3}\right) \pm i \sin\left(\frac{2\pi}{3}\right). \end{align}$$The fact that $\cos(2\pi/3)\pm i\sin(2\pi/3)$ are eigenvalues suggests that the matrix involves a rotation through the angle $2\pi/3$. To find the plane of rotation, we need the corresponding eigenvectors. Thus, we need to solve
$$ \begin{pmatrix} 0 & 0 & 1 \\ 1 & 0 & 0 \\ 0 & 1 & 0 \end{pmatrix} \begin{pmatrix} x\\y\\z \end{pmatrix} = \left(-\frac{1}{2}\pm\frac{\sqrt{3}}{2}i\right) \begin{pmatrix} x\\y\\z \end{pmatrix}. $$Taking the $+$ from the $\pm$, our system is equivalent to
$$\begin{align} z &= \left(-\frac{1}{2}+\frac{\sqrt{3}}{2}\right)x \\ x &= \left(-\frac{1}{2}+\frac{\sqrt{3}}{2}\right)y \\ y &= \left(-\frac{1}{2}+\frac{\sqrt{3}}{2}\right)z \end{align}.$$Allowing $z$ to be free and setting $z=1$, we obtain $x = -\frac{1}{2}-\frac{\sqrt{3}}{2} \text{ and } y = -\frac{1}{2}+\frac{\sqrt{3}}{2}$ so that an eigenvector is
$$\begin{pmatrix} -\frac{1}{2}-\frac{\sqrt{3}}{2} \\ -\frac{1}{2}+\frac{\sqrt{3}}{2} \\ 1 \end{pmatrix} = \begin{pmatrix} -1/2 \\ -1/2 \\ 1 \end{pmatrix} + \begin{pmatrix} -\sqrt{3}/{2} \\ \sqrt{3}/{2} \\ 0 \end{pmatrix}i. $$To interpret this, we could say that multiplication by $M$ fixes the line spanned by the vector $\langle 1,1,1 \rangle$ and rotates the plane spanned by the vectors $\langle -1/2,-1/2,1\rangle$ and $\langle -\sqrt{3}/{2}, \sqrt{3}/{2}, 0\rangle$.
Generally, the situation can be a bit more complicated. Given a matrix $M\in\mathbb R^{3\times3}$, with one real eigenpair $(\lambda_r,\vec{v}_r)$ and two complex conjugate eigenpairs $(\lambda_{\pm},\vec{v}_{\pm})$,
Here's a randomly generated example, like one that I'll ask you in the HW. We'll use the eigenstructure to deduce the geometric effect of the matrix
$$M = \frac{1}{3} \begin{pmatrix} 2 & -2 & 2 \\ -2 & -4 & -4 \\ 0 & 5 & -1 \end{pmatrix}.$$Here's some Python code (and a live link) to compute the eigensystem:
from scipy.linalg import eig
M = [[2/3, -2/3, 2/3],
[-2/3, -4/3, -4/3],
[0, 5/3, -1/3]]
evals,evecs = eig(M)
print(evals)
print('---')
print(evecs)
# Output (slightly trimmed):
[ 0.582]+0.j -0.79+1.186j -0.79-1.186j]
---
[[-0.95+0.j -0.37-0.065j -0.37+0.065j]
[ 0.14+0.j -0.20+0.52j -0.20-0.52j]
[ 0.26+0.j 0.73+0.j 0.73-0.j]]
The first part of the output tells us the eigenvalues:
[0.582+0.j -0.791+1.186j -0.791-1.186j]
The j
s represent the imaginary unit; thus j**2==-1
. Thus, we see one real eigenvalue, indicating a compression of about 0.58, and a pair of complex conjugate eigenvalues. As the magnitudes of those are larger than 1, we anticipate some expansion, as well as rotation.
The next portion of the output is a matrix whose columns are the corresponding eigenvectors:
[[-0.95+0.j -0.37-0.065j -0.37+0.065j]
[ 0.14+0.j -0.20+0.52j -0.20-0.52j]
[ 0.26+0.j 0.73+0.j 0.73-0.j]]
Note that the last two columns have non-zero imaginary parts and contain essentially the same information. We'll need to decompose one of those into real and imaginary parts.
The middle column decomposes as
$$ \begin{pmatrix} -0.37 - 0.065i \\ -0.2+0.52i \\ 0.73 \end{pmatrix} = \begin{pmatrix} -0.37\\ -0.2\\ 0.73 \end{pmatrix} + \begin{pmatrix} - 0.065 \\ 0.52 \\ 0 \end{pmatrix}i $$Those two vectors ($\langle -0.37, -0.2, 0.73 \rangle$ and $\langle - 0.065, 0.52, 0.73 \rangle$) span the subspace where the rotation happens.
The image below illustrates. The black vectors come from the eigenstructure of the matrixand the red vectors are the images of those black vectors.
As we move on, it will often be convenient to assume that a matrix is diagonalizable. Of course, not all matrices are diagonalizable but allowing the use of complex eigenvalues and eigenvectors greatly extends the class of matrices that are.
For example, a rotation matrix on $\mathbb R^2$ is certainly not diagonalizable since that would imply the existence of invariant 1D subspaces.
As an example, you may feel free to verify the following by direct computation:
$$ \begin{pmatrix} 0 & -1 \\ 1 & 0 \end{pmatrix} = \begin{pmatrix} i & -i \\ 1 & 1 \end{pmatrix} \begin{pmatrix} i & 0 \\ 0 & -i \end{pmatrix} \begin{pmatrix} i & -i \\ 1 & 1 \end{pmatrix}^{-1} $$The matrices $S$ and $D$ above are chosen using eigenvalues and eigenvectors exactly we did before - they just have complex components.