Eigenvalues and Eigenvectors

For a matrix transformation \( T \), a non-zero vector \( v\, (\neq 0) \) is called its eigenvector if \( T v = \lambda v \) for some scalar \( \lambda \). This means that applying the matrix transformation to the vector only scales the vector. The corresponding value of \( \lambda \) for \( v \) is an eigenvalue of \( T \).

The matrix transformation \(A\) acts on the eigenvector \(x\), scaling it by a factor of the eigenvalue \(\lambda\).

The matrix transformation \(A\) acts on the eigenvector \(x\), scaling it by a factor of the eigenvalue \(\lambda\). [1]

Contents

Matrix Transformation Example

To understand how matrix transformations act geometrically, it is important to understand their effects on individual vectors.

Consider the matrix transformation

This corresponds to the map \(T: \mathbb^2 \to \mathbb^2\) given by \((x,y) \mapsto (2x, 3y)\). The action of \(T\) is to scale the \(x\)-axis by a factor of 2 and the \(y\)-axis by a factor of 3. In particular, this is equivalent to requiring that \(T\) send \((1,0) \mapsto (2,0)\) and \((0,1) \mapsto (0,3)\). Assuming this, linearity of \(T\) implies that

\[\begin T(x,y) &= T\big((x,0) + (0,y)\big) \\ &= T(x,0) + T(0,y) \\ &= x \cdot T(1,0) + y \cdot T(0,1) \\ &= x \cdot (2,0) + y \cdot (0,3) \\ &= (2x, 3y). \end\]

Thus, it was possible to identify a basis \(\\) for \(\mathbb^2\) on which \(T\) acts very simply, by scaling. From this knowledge, one can deduce how \(T\) works on every vector in \(\mathbb^2\).

The natural question that sprouts out of this discussion is whether or not it is possible for all matrices to determine such a basis of eigenvectors. At the very least, when it is possible to do so, the matrix at hand may be considered especially well-behaved. For instance, the spectral theorem of linear algebra states that whenever \(A\) is a symmetric matrix, i.e. \(A = A^\), there is a basis consisting of eigenvectors for \(A\).

Computing Eigenvalues and Eigenvectors

It is not too difficult to compute eigenvalues and their corresponding eigenvectors when the matrix transformation at hand has a clear geometric interpretation. For examples, consider the diagonal matrix discussed above and the reflection matrix below:

Consider the reflection matrix transformation \( T = \begin -1 & 0 \\ 0 & 1 \end \) which reflects a vector across the \( y \)-axis. Find the eigenvectors and the corresponding eigenvalues of \( T \).

The vectors that get scaled after a reflection across the \( y \)-axis are either parallel to the \( y \)-axis, i.e. in the span of \((0,1)\), or parallel to the \(x\)-axis, i.e. in the span of \((1,0)\).

In the first case, the vector is not changed at all:

\[T(0,1) = (0,1) = 1\cdot (0,1),\]

and therefore the eigenvalue for \( (0,1) \) is \( 1 \).

In the second case, after applying the transformation, the length of the vector remains the same, but the direction reverses:

\[T(1,0) = (-1,0) = -1 \cdot (1,0),\]

and therefore the eigenvalue for \( (1,0) \) is \( -1 \). \(_\square\)

For arbitrary matrices, however, a more general procedure is necessary.

Let \(A\) be an \(n\)-by-\(n\) matrix, so that it corresponds to a transformation \(\mathbb^n \to \mathbb^n\). If \(\lambda\) is an eigenvalue for \(A\), then there is a vector \(v \in \mathbb^n\) such that \(Av = \lambda v\). Rearranging this equation shows that \((A - \lambda \cdot I)v = 0\), where \(I\) denotes the \(n\)-by-\(n\) identity matrix. This implies the null space of the matrix \(A-\lambda \cdot I\) is nonzero, so \(A-\lambda \cdot I\) has determinant zero.

Note that for every matrix \(A\) has 0 as an eigenvalue, with eigenvector \((0,0, \cdots, 0) \in \mathbb^n\). Generally, one is only concerned with the nonzero eigenvectors associated to an eigenvalue, so convention dictates that \(0\) is considered an eigenvalue of \(A\) only when the null space of \(A\) is nonzero \(\big(\)equivalently, when \(x\) divides \(p_ (x)\big).\)

In computations, the characteristic polynomial is extremely useful. To determine the eigenvalues of a matrix \(A\), one solves for the roots of \(p_ (x)\), and then checks if each root is an eigenvalue.

Connections with Trace and Determinant

Suppose \(A\) is a square matrix. Then

The proof of these properties requires the investigation of the characteristic polynomial of \(A\), which is found by taking the determinant of \((A - \lambda_)\). We have

Let \(_, _, . _\) be the roots of an \(n\)-order polynomial.

Since the eigenvalues are the roots of a matrix polynomial, we can match \(P(x)\) to \(\det(A - \lambda_)\). Therefore, it is clear that

Applications

In addition to their theoretical significance, eigenvalues and eigenvectors have important applications in various branches of applied mathematics, including signal processing, machine learning, and social network analysis.

For a simple application in network analysis, consider a network modeled by the finite connected graph \(G = (V, E)\). A computer will store the information of this graph in the form of an adjacency matrix. This is an \(n\)-by-\(n\) matrix \(A = \\>_<1\le i \le j \le n>\), where \(n = |V|\) and \(a_ = 1\) if vertices \(i\) and \(j\) have an edge between them, and \(a_ = 0\) otherwise.

The eigenvalues of \(A\) detect information about the structure of \(G\). For instance, let \(\lambda_>\) and \(\lambda_>\) denote the smallest and largest eigenvalues of \(A\), respectively. Then, \(G\) is bipartite if and only if \(\lambda_> = - \lambda_>\). Accordingly, a computer could determine whether or not \(G\) is bipartite by computing the eigenvalues of \(A\); this is far more computationally feasible than attempting to check \(G\) is bipartite by brute force.

References

  1. Lantonov, L. Eigenvalue-equation. Retrieved August 24, 2016, from https://en.wikipedia.org/wiki/Eigenvalues_and_eigenvectors#/media/File:Eigenvalue_equation.svg