Householder transformation

A Householder transformation of a vector ${\bf x}$ is its reflection (mirror image) with respect to a plane or hyperplane (as a mirror) through the origin with unit normal vector ${\bf v}$ ( $\vert\vert{\bf v}\vert\vert=1$):

$\displaystyle {\bf x}'={\bf x}-2 {\bf p}_{\bf v}({\bf x})
={\bf x}-2{\bf v}{\bf v}^T{\bf x}
=({\bf I}-2{\bf v}{\bf v}^T){\bf x}={\bf Px}$ (51)

where ${\bf p}_{\bf v}({\bf x})={\bf v}{\bf v}^T{\bf x}$ is the projection of ${\bf x}$ onto the unit vector ${\bf v}$, and ${\bf P}={\bf I}-2{\bf v}{\bf v}^T$ is the Householder matrix that converts ${\bf x}$ to its reflection ${\bf x}'$. Given the desired reflection ${\bf x}'$, we can get the normal direction of the reflection plane ${\bf u}={\bf x}-{\bf x}'$, and then normalize it to get ${\bf v}={\bf u}/\vert\vert{\bf u}\vert\vert$. Although the Householder transformation is expressed as a linear transformation ${\bf x}'={\bf Px}$, its computational complexity is $O(N)$, instead of $O(N^2)$ for a general matrix multiplication ${\bf Ax}$.

Householder3.png

The Householder matrix ${\bf P}$ is symmetric

$\displaystyle {\bf P}^T=({\bf I}-2{\bf v}{\bf v}^T)^T={\bf I}-2{\bf v}{\bf v}^T={\bf P}$ (52)

and orthogonal:
$\displaystyle {\bf P}^T{\bf P}={\bf P}{\bf P}$ $\displaystyle =$ $\displaystyle ({\bf I}-2{\bf v}{\bf v}^T)({\bf I}-2{\bf v}{\bf v}^T)
={\bf I}-2{\bf v}{\bf v}^T-2{\bf v}{\bf v}^T
+4{\bf v}{\bf v}^T{\bf v}{\bf v}^T$  
  $\displaystyle =$ $\displaystyle {\bf I}-4{\bf v}{\bf v}^T+4{\bf v}{\bf v}^T={\bf I}$ (53)

i.e., ${\bf P}={\bf P}^T={\bf P}^{-1}$. As vector norm is preserved by an orthogonal transform, we have $\vert\vert{\bf x}\vert\vert=\vert\vert{\bf Px}\vert\vert=\vert\vert{\bf x}'\vert\vert$.

We can verify that ${\bf P}$ indeed converts the given ${\bf x}$ into the desired ${\bf x}'$ if ${\bf u}={\bf x}-{\bf x}'$, or $-{\bf x}'$ if ${\bf u}={\bf x}+{\bf x}'$:

$\displaystyle {\bf P}{\bf x}$ $\displaystyle =$ $\displaystyle \left( {\bf I}-\frac{2{\bf u}({\bf x}\pm{\bf x}')^T}
{({\bf x}\pm...
...m{\bf x}'^T{\bf x})}
{{\bf x}^T{\bf x}\pm2{\bf x}^T{\bf x}'+{\bf x}'^T{\bf x}'}$  
  $\displaystyle =$ $\displaystyle {\bf x}-\frac{{\bf u}({\bf x}^T{\bf x}\pm{\bf x}'^T{\bf x})}
{({\...
...pm{\bf x}'^T{\bf x})}
={\bf x}-{\bf u}={\bf x}-({\bf x}\pm{\bf x}')=\mp{\bf x}'$ (54)

Householder2b.png

The Householder transformation is typically used to convert a given vector ${\bf x}=[x_1,\cdots,x_N]^T$ into a vector with all but the first one of its components converted to zero:

$\displaystyle {\bf x}'=[x'_1,0,\cdots,0]^T=\vert\vert{\bf x}\vert\vert{\bf e}_1$ (55)

where ${\bf e}_1=[1,0,\cdots,0]^T$ is the first standard basis vector. In this case, the normal vector needs to be:

$\displaystyle {\bf u}={\bf x}\pm{\bf x}' ={\bf x}\pm\vert\vert\;{\bf x}\vert\vert{\bf e}_1$ (56)

We note that the reflection ${\bf x}'=\vert\vert{\bf x}\vert\vert\;[1,0,\cdots,0]^T$ is always a real vector. However, when the given vector ${\bf x}$ is complex, the reflection needs to be kept complex as well, instead of being forced to be real. In this case, we should use ${\bf x}'=\vert\vert{\bf x}\vert\vert\;[x_1/\vert x_1\vert,0,\cdots,0]^T$, where $x_1/\vert x_1\vert$ is a normalized complex value based on the first entry of the given vector ${\bf x}$.

By applying a sequence of Householder transformations to the rows and columns of a given square matrix ${\bf A}$, we can convert it into certain desired form, such as a tridiagonal matrix if ${\bf A}$ is symmetric, or a Hessenberg matrix if ${\bf A}$ is non-symmetric. This process is actually a similarity transformation ${\bf Q}{\bf A}{\bf Q}^{-1}$, which preserves the eigenvalues of ${\bf A}$. Some subsequent algorithm can be further applied to solve the eigenvalue problem an arbitrary matrix ${\bf A}$.

We first consider converting a symmetric matrix ${\bf A}={\bf A}^T$ to a tridiagonal matrix. To start, we construct an $N\times N$ square matrix:

$\displaystyle {\bf Q}_1=\left[\begin{array}{c\vert ccc}
1&0&\cdots&0\\ \hline 0& & & \\ \vdots & &{\bf P}_{N-1} & \\
0& & & \end{array}\right]$ (57)

where ${\bf P}_{N-1}$ is an $N-1$ dimensional Householder matrix that converts $[a_{21},\cdots,a_{N1}]^T$ containing the last $N-1$ elements of the first column of ${\bf A}$ into $[a'_{21},0,\cdots,0]^T$. Note that ${\bf Q}_1={\bf Q}_1^T={\bf Q}_1^{-1}$ is also symmetric and orthogonal. Pre-multiplying ${\bf Q}_1$ to ${\bf A}$ we get:
$\displaystyle {\bf Q}_1{\bf A}$ $\displaystyle =$ $\displaystyle \left[\begin{array}{c\vert ccc}
1&0&\cdots&0\\ \hline 0& & & \\ \...
...ine
a_{21}& & & \\ \vdots & &{\bf A}_{N-1} & \\ a_{N1} & & & \end{array}\right]$  
  $\displaystyle =$ $\displaystyle \left[\begin{array}{c\vert ccc}a_{11}&a_{12}&\cdots&a_{1N}\\ \hli...
...\\ a_{N1}\end{array}\right)
& & {\bf P}_{N-1}{\bf A}_{N-1} & \end{array}\right]$  
  $\displaystyle =$ $\displaystyle \left[\begin{array}{c\vert ccc}a_{11}&a_{12}&\cdots&a_{1N}\\ \hli...
...0 & & & \\ \vdots & &{\bf P}_{N-1}{\bf A}_{N-1} &
\\ 0 & & & \end{array}\right]$ (58)

where ${\bf A}_{N-1}$ is an $N-1$ dimensional submatrix containing the lower-right $N-1$ by $N-1$ elements of ${\bf A}$. We see that the last $N-2$ components of the first column of the resulting matrix ${\bf Q}_1{\bf A}$ are zero. We next post-multiply ${\bf Q}_1$ to modify the first row:
$\displaystyle {\bf A}'={\bf Q}_1{\bf A}{\bf Q}_1$ $\displaystyle =$ \begin{displaymath}\left[\begin{array}{c\vert ccc}
a_{11}&a_{12}&\cdots&a_{1N}\\...
...{\bf P}_{N-1}&\\
\vdots & & & \\ 0& & & \\
\end{array}\right]\end{displaymath}  
  $\displaystyle =$ $\displaystyle \left[\begin{array}{c\vert ccc}a_{11}& &
[a_{12},\cdots,a_{1N}]{\...
...dots & &{\bf P}_{N-1}{\bf A}_{N-1}{\bf P}_{N-1} &
\\ 0 & & & \end{array}\right]$  
  $\displaystyle =$ $\displaystyle \left[\begin{array}{c\vert cccc}a_{11}&a'_{12}&0&\cdots&0\\ \hlin...
...ts &&&{\bf P}_{N-1}{\bf A}_{N-1}{\bf P}_{N-1} &
\\ 0 & & & & \end{array}\right]$ (59)

We see that the last $N-2$ components of the first row and column of ${\bf A}'$ become zero.

This process is then repeated for the lower-right submatrix ${\bf P}_{N-1}{\bf A}_{N-1}{\bf P}_{N-1}$ of ${\bf A}'$. Pre and post multiplying ${\bf A}'$ by

$\displaystyle {\bf Q}_2=\left[\begin{array}{cc\vert ccc}
1&0&0&\cdots&0\\ 0&1&0...
...ne
0&0& & & \\ \vdots &\vdots& &{\bf P}_{N-2} & \\
0&0& & & \end{array}\right]$ (60)

we get

$\displaystyle {\bf A}''={\bf Q}_2{\bf A}'{\bf Q}_2={\bf Q}_2{\bf Q}_1{\bf A}{\b...
...{\bf P}_{N-2}{\bf A}_{N-2}{\bf P}_{N-2}& \\
0 &0& & & & \\
\end{array}\right]$ (61)

We see that the last $N-3$ components of the second row and column of ${\bf A}''$ become zero. This process is repeated until after $N-1$ iterations the N-D matrix ${\bf A}$ is converted into a tridiagonal matrix, with all entries $a_{ij}$ equal to zero except those along the sub-diagonal ($i=j+1$), the superdiagonal ($i=j-1$) as well as the main diagonal ($i=j$):

$\displaystyle {\bf Q}_{N-1}\cdots{\bf Q}_1{\bf A}{\bf Q}_1\cdots{\bf Q}_{N-1}
={\bf QAQ}^T={\bf QAQ}^{-1}$ (62)

where we have defined ${\bf Q}={\bf Q}_{N-1}\cdots{\bf Q}_1$. Note that ${\bf Q}$ is not symmetric (a product of symmetric matrices is not symmetric), but it is orthogonal:
$\displaystyle {\bf Q}^T$ $\displaystyle =$ $\displaystyle ({\bf Q}_{N-1}\cdots{\bf Q}_1)^T
={\bf Q}_1^T\cdots{\bf Q}_{N-1}^T$  
  $\displaystyle =$ $\displaystyle {\bf Q}_1^{-1}\cdots{\bf Q}_{N-1}^{-1}
=({\bf Q}_{N-1}\cdots{\bf Q}_1)^T={\bf Q}^{-1}$ (63)

The resulting matrix ${\bf QAQ}^{-1}$, a similarity transformation of ${\bf A}$, has the same eigenvalues as ${\bf A}$.

Based on the example tridiagonalization above, we make a few notes

Examples: Consider two matrices, one symmetric, one non-symmetric: