next up previous
Next: About this document ... Up: VectorSpace Previous: VectorSpace

Vector Space and Orthogonal Transform

Definition: A vector space is a set $V$ with two operations of addition and scalar multiplication defined for its members, referred to as vectors.

  1. Vector addition maps any two vectors ${\bf x}, {\bf y} \in V$ to another vector ${\bf x}+{\bf y} \in V$ satisfying the following properties:
  2. Scalar multiplication maps a vector ${\bf x}\in V$ and a real or complex scalar $a\in \mathbb{C}$ to another vector $a{\bf x} \in V$ with the following properties:

Definition: An inner product on a vector space $V$ is a function that maps two vectors ${\bf x}, {\bf y} \in V$ to a scalar $\langle {\bf x},{\bf y}\rangle \in
\mathbb{C}$ and satisfies the following conditions:

Definition: A vector space with inner product defined is called an inner product space.

Definition: When the inner product is defined, $\mathbb{C}^N$ is called a unitary space and $\mathbb{R}^N$ is called a Euclidean space.

Examples

The concept of inner product is of essential importance based on which a whole set of other important concepts can be defined.

Definition: If the inner product of two vectors ${\bf x}$ and ${\bf y}$ is zero, $\langle {\bf x}, {\bf y}\rangle=0$, they are orthogonal (perpendicular) to each other, denoted by ${\bf x}\; \bot\; {\bf y}$.

Definition: The norm (or length) of a vector ${\bf x}\in V$ is defined as

\begin{displaymath}\vert\vert{\bf x}\vert\vert=\sqrt{\langle {\bf x},{\bf x}\ran...
... \vert\vert{\bf x}\vert\vert^2=\langle {\bf x},{\bf x}\rangle. \end{displaymath}

The norm $\vert\vert{\bf x}\vert\vert$ is non-negative and it is zero if and only if ${\bf x}={\bf0}$. In particular, if $\vert\vert{\bf x}\vert\vert=1$, then it is said to be normalized and becomes a unit vector. Any vector can be normalized when divided by its own norm: ${\bf x}/\vert\vert{\bf x}\vert\vert$. The vector norm squared $\vert\vert{\bf x}\vert\vert^2=\langle {\bf x},{\bf x}\rangle$ can be considered as the energy of the vector.

Example: In an $N$-D unitary space, the p-norm of a vector ${\bf x}=[x[1],\ldots,x[N]]^\textrm{T} \in \mathbb{C}^N$ is

\begin{displaymath}\vert\vert{\bf x}\vert\vert _p=\left[\sum_{n=1}^N \vert x[n]\vert^p\right]^{1/p}. \end{displaymath}

The concept of $N$-D unitary (or Euclidean) space can be generalized to an infinite-dimensional space, in which case the range of the summation will cover all real integers ${\mathbb Z}$ in the entire real axis $-\infty<n<\infty$. This norm exists only if the summation converges to a finite value; i.e., the vector ${\bf x}$ is an energy signal with finite energy:

\begin{displaymath}\sum_{n=-\infty}^\infty \vert x[n]\vert^2 < \infty. \end{displaymath}

All such vectors ${\bf x}$ satisfying the above are square-summable and form the vector space denoted by $l^2({\mathbb Z})$.

Similarly, in a function space, the norm of a function vector ${\bf x}=x(t)$ is defined as

\begin{displaymath}\vert\vert{\bf x}\vert\vert=\left[ \int_a^b x(t)\overline{x(t...
...{1/2}
=\left[ \int_a^b \vert x(t)\vert^2\; \;dt \right]^{1/2}, \end{displaymath}

where the lower and upper integral limits $a<b$ are two real numbers, which may be extended to all real values $\mathbb{R}$ in the entire real axis $-\infty<t<\infty$. This norm exists only if the integral converges to a finite value; i.e., $x(t)$ is an energy signal containing finite energy:

\begin{displaymath}\int_{-\infty}^\infty \vert x(t) \vert^2 \;dt < \infty. \end{displaymath}

All such functions $x(t)$ satisfying the above are square-integrable, and they form a function space denoted by ${\cal L}^2(\mathbb{R})$.

Definition: In a unitary space $\mathbb{C}^N$, the p-norm distance between two vectors ${\bf x}$ and ${\bf y}$ is defined as the p-norm of the difference ${\bf x}-{\bf y}$:

\begin{displaymath}d_p({\bf x},{\bf y})=\left( \sum_{n=1}^N \vert x[n]-y[n]\vert^p\right)^{1/p}. \end{displaymath}

In a function space, the p-norm distance between two functions $x(t)$ and $y(t)$ is similarly defined as

\begin{displaymath}d_p(x(t),y(t))=\left( \int_a^b \vert x(t)-y(t)\vert^p \; \;dt\right)^{1/p}. \end{displaymath}

In particular, when $p=2$, we have

\begin{displaymath}d_2(x(t),y(t))=\vert\vert x(t)-y(t)\vert\vert=\left( \int_a^b \vert x(t)-y(t)\vert^2 \;dt\right)^{1/2}.
\end{displaymath}

Definition: A vector space V of all linear combinations of a set of vectors ${\bf b}_k\in V,\;(k=1,\ldots,N)$ is called the linear span of the vectors:

\begin{displaymath}W=span({\bf b}_1,\ldots,{\bf b}_N)
=\left\{ \sum_{k=1}^N c_k {\bf b}_k\;\;\big\vert\;\; c_k \in \mathbb{C} \right\}.
\end{displaymath}

Definition: A set of linearly independent vectors that spans a vector space is called a basis of the space. It these vectors are unitary/orthogonal and normalized, they form an orthonormal basis.

Any vector ${\bf x}\in \mathbb{C}^N$ can be uniquely expressed as a linear combination of some $N$ basis vectors ${\bf b}_k$:

\begin{displaymath}{\bf x}=\sum_{k=1}^N c_k {\bf b}_k. \end{displaymath}

The concept of a finite $N$-D space spanned by a basis composed of $N$ discrete (countable) linearly independent vectors can be generalized to a vector space V spanned by a basis composed of a family of uncountably infinite vectors ${\bf b}(f)$. Any vector ${\bf x}\in V$ in the space can be expressed as a linear combination, an integral, of these basis vectors:

\begin{displaymath}{\bf x}=\int_a^b c(f) {\bf b}(f) \;df. \end{displaymath}

Theorem: Let ${\bf x}$ and ${\bf y}$ be any two vectors in a vector space $V$ spanned by a set of complete orthonormal (orthogonal and normalized) basis vectors $\{ {\bf u}_k \}$ satisfying

\begin{displaymath}\langle {\bf u}_k,{\bf u}_l\rangle=\delta[k-l]. \end{displaymath}

Then we have
  1. Series expansion:

    \begin{displaymath}{\bf x}=\sum_k \langle {\bf x},{\bf u}_k\rangle{\bf u}_k
=\s...
...;\;\mbox{where}\;\;\;\;
c_k=\langle {\bf x},{\bf u}_k\rangle. \end{displaymath}


    \begin{displaymath}
{\bf y}=\sum_k \langle {\bf y},{\bf u}_k\rangle{\bf u}_k
=...
...;\mbox{where}\;\;\;\;
d_k=\langle {\bf y},{\bf u}_k\rangle.
\end{displaymath}

  2. Plancherel theorem:

    \begin{displaymath}\langle {\bf x},{\bf y}\rangle=\sum_k \langle
{\bf x},{\bf ...
...;\;\mbox{where}\;\;\;\;
d_k=\langle {\bf y},{\bf u}_k\rangle. \end{displaymath}

  3. Parseval's theorem:

    \begin{displaymath}\langle {\bf x},{\bf x}\rangle=\vert\vert{\bf x}\vert\vert^2=...
...angle {\bf x},{\bf u}_k\rangle\vert^2=\sum_k \vert c_k\vert^2. \end{displaymath}

Proof: Taking an inner product with ${\bf u}_l$ on both sides of the first equation we get

\begin{displaymath}\langle {\bf x},{\bf u}_l\rangle=\left\langle \sum_k c_k {\bf...
...\langle {\bf u}_k,{\bf u}_l\rangle=\sum_k c_k \delta[k-l]=c_l.
\end{displaymath}

Here, ${\bf x}$ is expressed as the vector sum of its projections ${\bf p}_{{\bf u}_k}({\bf x})=\langle {\bf x},{\bf u}_k\rangle {\bf u}_k$ onto each of the unit basis vectors ${\bf u}_k$ and the scalar coefficient $c_k=\langle {\bf x},{\bf u}_k\rangle$ is the norm of the projection.

Example: Space $\mathbb{C}^N$ can be spanned by $N$ orthonormal vectors $\{ {\bf u}_1,\ldots,{\bf u}_N\}$, where the $k$th basis vector is ${\bf u}_k=[u[1,k],\ldots,u[N,k]]^\textrm{T}$, that satisfy:

\begin{displaymath}\langle {\bf u}_k,{\bf u}_l\rangle={\bf u}_k^\textrm{T} \over...
...{\bf u}}_l
=\sum_{n=1}^N u[n,k]\overline{u}[n,l]=\delta[k-l]. \end{displaymath}

Any vector ${\bf x}=[x[1],\ldots,x[N]]^\textrm{T} \in \mathbb{C}^N$ can be expressed as

\begin{displaymath}{\bf x}=\sum_{k=1}^N c_k {\bf u}_k=[{\bf u}_1,\ldots,{\bf u}_...
...y}{c}c[1] \vdots  c[N]\end{array}\right]
={\bf U}{\bf c},
\end{displaymath}

where ${\bf c}=[c[1],\ldots,c[N]]^\textrm{T}$ and

\begin{displaymath}{\bf U}=[{\bf u}_1,\ldots,{\bf u}_N]=\left[ \begin{array}{ccc...
...ts & \vdots \\
u[N,1] & \ldots & u[N,N] \end{array} \right].
\end{displaymath}

As the column (and row) vectors in ${\bf U}$ are orthogonal, it is a unitary matrix that satisfies ${\bf U}^{-1}={\bf U}^*$; i.e., ${\bf U}{\bf U}^*={\bf U}^*{\bf U}={\bf I}$ To find the coefficient vector ${\bf c}$, we pre-multiply ${\bf U}^{-1}={\bf U}^*$ on both sides of the previous equation and get:

\begin{displaymath}{\bf U}^*{\bf x}={\bf U}^{-1}{\bf x}={\bf U}^{-1}{\bf U}{\bf c}={\bf c}. \end{displaymath}

This two equations can be rewritten as a pair of transforms:

\begin{displaymath}\left\{ \begin{array}{l}
{\bf c}={\bf U}^*{\bf x}={\bf U}^{-1}{\bf x} \\
{\bf x}={\bf U}{\bf c}
\end{array} \right.. \end{displaymath}

We see that the norm of ${\bf x}$ is conserved (Parseval's identity):

\begin{displaymath}\vert\vert{\bf x}\vert\vert^2=\langle {\bf x},{\bf x}\rangle=...
...=\langle {\bf c},{\bf c}\rangle=\vert\vert{\bf c}\vert\vert^2. \end{displaymath}

The second equation in the transform pair can also be written in component form as

\begin{displaymath}x[n]=\sum_{k=1}^N c_k u[k,n],\;\;\;\;\;\;\;\;\;n=1,\ldots,N. \end{displaymath}

Obviously, the $N$ coefficients $c_k$ ($k=1,\ldots,N$) can be obtained with computational complexity $O(N^2)$.

Example: In ${\cal L}^2$ space composed of all square-integrable functions defined over $a<t<b$, spanned by a set of orthonormal basis functions $u_k(t)$ satisfying:

\begin{displaymath}
\langle u_k(t),u_l(t)\rangle=\int_a^b u_k(t)\overline{u}_l(t) \;dt=\delta[k-l].
\end{displaymath}

Any $x(t)$ in the space can be written as

\begin{displaymath}x(t)=\sum_k c_k u_k(t). \end{displaymath}

Taking an inner product with $u_l(t)$ on both sides, we get

\begin{displaymath}\langle x(t),u_l(t)\rangle=\sum_k c_k\langle u_k(t), u_l(t)\rangle=\sum_k c_k\delta[k-l]=c_l; \end{displaymath}

i.e.,

\begin{displaymath}c_k=\langle x(t), u_k(t)\rangle=\int_a^b x(t)\overline{u}_k(t) \;dt. \end{displaymath}

which is the projection of $x(t)$ onto the unit basis function $\phi_k(t)$. Again we can easily get:

\begin{displaymath}\vert\vert x(t)\vert\vert^2=\langle x(t),x(t)\rangle=\int_a^b...
...t) \;dt=\sum_k \vert c_k\vert^2=\vert\vert{\bf c}\vert\vert^2.
\end{displaymath}

The Fourier transforms

Consider the following four Fourier bases that span four different types of vector spaces for signals that are either continuous or discrete, of finite or infinite duration.

Definition: A linear transformation $U: V\rightarrow W$ is a unitary transformation if it conserves inner products:

\begin{displaymath}\langle {\bf x},{\bf y}\rangle=\langle U{\bf x},U{\bf y}\rangle. \end{displaymath}

In particular, if the vectors are real with symmetric inner product $\langle {\bf x},{\bf y}\rangle=\langle {\bf y},{\bf x}\rangle$, then $U$ is an orthogonal transformation.

A unitary transformation also conserves any measurement based on the inner product, such as the norm of a vector, the distance and angle between two vectors, and the projection of one vector on another. Also, if in particular ${\bf x}={\bf y}$, we have

\begin{displaymath}
\langle {\bf x},{\bf x}\rangle=\vert\vert{\bm f}\vert\vert^...
...gle U{\bf x},U{\bf x}\rangle=\vert\vert U{\bf x}\vert\vert^2;
\end{displaymath}

i.e., the unitary transformation conserves the vector norm (length). This is Parseval's identity for a generic unitary transformation $U{\bm x}$. Owing to this property, a unitary operation $R: V \rightarrow V$ can be intuitively interpreted as a rotation in space $V$.

Definition A matrix ${\bf U}$ is unitary if it conserves inner products:

\begin{displaymath}\langle {\bf U}{\bf x},{\bf U}{\bf y}\rangle=\langle {\bf x},{\bf y}\rangle.
\end{displaymath}

Theorem: A matrix ${\bf U}$ is unitary if and only if ${\bf U}^*{\bf U}={\bf I}$; i.e., the following two statements are equivalent:

$\displaystyle (a)$   $\displaystyle \langle {\bf U}{\bf x},{\bf U}{\bf y}\rangle=\langle {\bf x},{\bf y}\rangle$  
$\displaystyle (b)$   $\displaystyle {\bf U}^*{\bf U}={\bf U}{\bf U}^*={\bf I};\;\;\;\;\;
\mbox{i.e.,}\;\;\;\;\;\;{\bf U}^{-1}={\bf U}^*.$  

When ${\bf U}=\overline{{\bf U}}$ is real, it is an orthogonal matrix.

A unitary matrix ${\bf U}$ has the following properties:

The last property indicates that the column (row) vectors $\{ {\bf u}_k \}$ form an orthogonal basis that spans $\mathbb{C}^N$.

The identity matrix ${\bf I}=[{\bf e}_1,\ldots,{\bf e}_N]$ is a special orthogonal matrix as its columns (or rows) are orthonormal:

\begin{displaymath}\langle {\bf e}_k,{\bf e}_l\rangle=\delta[k-l]. \end{displaymath}

where the kth vector ${\bf e}_k=[0,\cdots,0,1,0,\cdots,0]^T$ has all zero elements except the kth one being 1. These vectors form the standard basis of $\mathbb{C}^N$. Any vector ${\bf x}=[x[1],\ldots,x[N]]^\textrm{T} \in \mathbb{C}^N$ can be represented by the standard basis as

\begin{displaymath}{\bf x}=\left[\begin{array}{c}x[1] \vdots x[N]\end{array}...
...ray}{c}x[1] \vdots x[N]\end{array}\right]
={\bf I}{\bf x}
\end{displaymath}

The vector can also be represented by any other orthonormal basis ${\bf U}=[{\bf u}_1,\ldots,{\bf u}_N]$ as

\begin{displaymath}{\bf x}={\bf I}{\bf x}={\bf U} {\bf U}^*{\bf x}={\bf U} {\bf ...
...\vdots c[N]\end{array}\right]
=\sum_{k=1}^N c[k] {\bf u}_k,
\end{displaymath}

where we have defined

\begin{displaymath}{\bf c}=\left[\begin{array}{c}c[1] \vdots c[N]\end{array}...
...;\;
c[k]={\bf u}_k^*{\bf x}=\langle {\bf x},{\bf u}_k\rangle.
\end{displaymath}

Combining the two equations we get

\begin{displaymath}\left\{ \begin{array}{l} {\bf c}={\bf U}^*{\bf x},\\
{\bf x}={\bf U}{\bf c}.
\end{array}\right.
\end{displaymath}

This is the generalized Fourier transform, by which a vector ${\bf x}$ is rotated to become another vector ${\bf c}$.

rotation.gif

This result can be extended to the continuous transformation for signal vectors in the form of continuous functions. In general, corresponding to any given unitary transformation $U$, a signal vector ${\bf x}\in V$ can be alternatively represented by a coefficient vector ${\bf c}=U^*{\bf x}$ (where ${\bf c}$ can be either a set of discrete coefficients $c_k$ or a continuous function $c(f)$). The original signal vector ${\bf x}$ can always be reconstructed from ${\bf c}$ by applying $U$ on both sides of ${\bf c}=U^*{\bf x}$ to get $U{\bf c}
=UU^*{\bf x}=I{\bf x}={\bf x}$; i.e., we get a unitary transform pair in the most general form:

\begin{displaymath}\left\{ \begin{array}{l} {\bf c}=U^*{\bf x},\\
{\bf x}=U{\bf c}.
\end{array}\right. \end{displaymath}

The first equation is the forward transform that maps the signal vector ${\bf x}$ to a coefficient vector ${\bf c}$, while the second equation is the inverse transform by which the signal is reconstructed.
next up previous
Next: About this document ... Up: VectorSpace Previous: VectorSpace
Ruye Wang 2013-10-16