<< Chapter < Page
  Signal theory   Page 1 / 1
Chapter >> Page >
Introduces the Karhuenen-Loeve transform, with applications.

Define the random vector

X = x 1 x 2 x N

with mean zero and covariance matrix R X = E [ X X * ] ; this matrix is symmetric and positive semidefinite.

Lemma 1 Every eigenvalue of R X is real and non-negative.

Let e be an eigenvector of R X with eigenvalue λ .

λ e 2 = λ e , e = λ e , e = A e , e 0 .

The last statement falls out by the definite of positive semi-definite. We have λ e 2 0 . Since e 2 0 , it follows that λ 0 , i.e. all the eigenvalues are non-negative. The eigenvectors of the matrix R X provide an orthonormal basis { φ 1 , φ 2 , . . . φ N } , which can be collected into an orthonormal basis matrix φ = [ φ 1 φ 2 ... φ N ] . Then let y = φ * x . We have:

R Y = E [ y y * ] = E [ φ * x x * φ ] = φ * E [ x x * ] φ = φ * R X φ .

Let us look at the adjoint of φ * R X :

φ * R X * = R X φ = R X [ φ 1 φ 2 ... φ N ] = [ λ 1 φ 1 λ 2 φ 2 ... λ N φ N ] .

If we take the adjoint again, we get

φ * R X = φ 1 * λ 1 φ 2 * λ 2 φ N * λ N .

Going back to our derivation of R Y :

R Y = φ * R X φ = φ 1 * λ 1 φ N * λ N φ 1 φ N = λ 1 φ 1 , φ 1 λ 1 φ 1 , φ 2 λ 1 φ 1 , φ N λ 2 φ 2 , φ 1 λ 2 φ 2 , φ 2 λ 2 φ 2 , φ N λ N φ N , φ 1 λ N φ N , φ 2 λ N φ N , φ N = λ 1 0 0 0 λ 2 0 0 0 λ N

The matrix φ is known as the KLT matrix defined by R X . The transformation given by the KLT matrix provides a set of random variables y i = φ i , x that are uncorrelated.

Example 1 (Whitening Filter) For a random vector X , R X has positive eigenvalues. Let us write R Y - 1 / 2 = diag ( λ 1 - 1 / 2 , . . . λ N - 1 / 2 ) and z = R Y - 1 / 2 y where y = φ * x . We have

R Z = E [ z z * ] = E [ R Y - 1 / 2 y y * R Y - 1 / 2 ] = R Y - 1 / 2 E [ y y * ] R Y - 1 / 2 = R Y - 1 / 2 R Y R Y - 1 / 2 = I .

The matrix R Y - 1 / 2 φ * is known as a "whitening filter", as it maps an arbitrary random vector x to a “white Gaussian noise” vector z .

Example 2 (Transform Coding) Let U : C n C n is a unitary operator. Assume we have a signal x C n that we want to send it through a channel by only sending k numbers or "items", where k < n ; in words, we wish to compress the signal x . The block diagram for the compression/transmission system is given in [link] .

Block diagram for a transform coding system.

We want to minimize E [ x - x ^ ] given k by choosing the optimal transformation U . We know y = U * x which implies x = U y since U is unitary. Therefore,

x - x ^ = U y - U y ^ = U ( y - y ^ ) = y - y ^ .

This means that we can minimize y - y ^ in place of x - x ^ . For simplicity, we choose a basic means of compression that preserves only the first k entries of y :

y i ^ = y i , if i = 1 , 2 , . . . k , 0 , if i = k + 1 , k + 2 , . . . n .

We then have E [ x - x ^ 2 ] = E [ y - y ^ 2 ] = E [ i = k + 1 n | y i | 2 ] = i = k + 1 n E [ | y i | 2 ] . Therefore,

min x ^ ( E [ x - x ^ 2 ] ) = min U i = k + 1 n E [ | y i | 2 ] = min U i = k + 1 n E [ | x , u i | 2 ] , = min U i = k + 1 n E [ u i T x x T u i ] = min U i = k + 1 n u i T E [ x x T ] u i , = min U i = k + 1 n u i T R X u i .

It turns out that the choice of transform basis U that minimizes this amount is provided by the eigendecomposition of R X , as specified by the following theorem.

Theorem 1 Let X be a length-n random vector with covariance matrix R X = E [ x x * ] that has eigenvalues λ 1 λ 2 λ 3 ... λ n 0 and matching eigenvectors φ 1 , φ 2 , ... φ N . Let x M be the orthogonal projection of x onto a subspace M of dimension k . Then

E [ | | x - x M | | 2 ] i = k + 1 M λ i ,

with equality if M = span ( { φ 1 , φ 2 , ... φ k } ) .

From equation [link] , we have

min M E [ | | x - x M | | 2 ] = min U i = k + 1 n u i T R X u i = min U i = k + 1 n u i T Φ Λ Φ T u i ,

where R X = Φ Λ Φ T is the eigendecomposition of R X . Now, since

Φ T u i = u i , φ 1 u i , φ 2 u i , φ n and Λ Φ T u i = u i , φ 1 λ 1 u i , φ 2 λ 2 u i , φ n λ n ,

we have that u i T Φ Λ Φ T u i = ( Φ T u i ) T Λ Φ T u i = i = j n | u i , φ j | 2 λ j . Plugging this into [link] , we have

min M E [ | | x - x M | | 2 ] = min U i = k + 1 n i = j n | u i , φ j | 2 λ j = min U i = j n λ j i = k + 1 n | u i , φ j | 2 .

Now, denote α j = i = k + 1 n | u i , φ j | 2 , and see that

j = 1 n α j = j = 1 n i = k + 1 n | u i , φ j | 2 = i = k + 1 n j = 1 n | u i , φ j | 2 = i = k + 1 n | | u i | | 2 = n - k ,

as all u i are unit-norm. Now, we have that

min M E [ | | x - x M | | 2 ] = min U i = j n λ j α j = min U j = 1 k λ j α j - j = k + 1 n λ j ( 1 - α j ) + j = k + 1 n λ j .

Since the λ k are monotonically decreasing, we have that

min M E [ | | x - x M | | 2 ] min U j = 1 k λ k α j - j = k + 1 n λ k ( 1 - α j ) + j = k + 1 n λ j , min U λ k j = 1 k α j - j = k + 1 n 1 + j = k + 1 n α j + j = k + 1 n λ j , min U λ k j = 1 n α j - ( n - k ) + j = k + 1 n λ j , min U λ k j = 1 k α j - λ k ( n - k ) + j = k + 1 n λ j , min U λ k ( n - k ) - λ k ( n - k ) + j = k + 1 n λ j , min U j = k + 1 n λ j .

If we set M = span { φ 1 , φ 2 , . . . φ k } ) , (i.e., U = Φ ) then it is easy to check that

E [ | | x - x M | | 2 ] = j = k + 1 n λ j ,

proving the theorem.

Example 3 (Transform Coding) Transform coding is a common scheme for data compression that leverages the Karhuenen-Loève transform. Examples include JPEG and MP3. In particular, JPEG can be broadly described as follows:

  1. Take the image x and create tiles of size 8 × 8 . We assume that the tiles are draws from a random variable X , i.e., the tiles x 1 , x 2 ... R 64 with R X = 1 n i = 1 n x i x i T
  2. Compute the KLT of the tile random variable X from R X by obtaining its eigendecomposition R X = Φ Λ Φ T .
  3. Compute KLT coefficients for each block as c i = Φ T x i .
  4. Pick as many coefficients of c i as allowed by communications or storage constraints; save them as the compressed image.
  5. Load saved coefficients and append zeros to build coefficient vector c ^ i .
  6. Run inverse KLT to obtain the decompressed tiles x ^ i = Φ c ^ i .
  7. Reassemble the image from the decompressed tiles.

In practice, it is not desirable to recompute the KLT for each individual image. Thus, the JPEG algorithm employs the discrete cosine transform (DCT). It turns out that the DCT is a good approximation of the KLT for tiles of natural images. Additionally, instead of selecting a subset of the coefficients, they are quantized to varying quality/error according to their index and the total amount of bits available.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Signal theory. OpenStax CNX. Oct 18, 2013 Download for free at http://legacy.cnx.org/content/col11542/1.3
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Signal theory' conversation and receive update notifications?

Ask