<< Chapter < Page Chapter >> Page >

Continuing from last time, we have a signal x R N , an n × N measurement matrix Φ , and y = Φ x R n is the information we draw from x .

We consider decoders Δ mapping R n R N . We have been discussing whether there exists a decoder withcertain properties. So for this discussion (about information preservation), we can just think about optimal decoding.

While the previous result on sparse signal recovery is interesting, it is not very robust. What happens if our signal does not have support k ? Can we still obtain meaningful results? The answer is yes. To do so we introduce more general input signal classes K X that allows fully supported signals. For example, we will consider the signal class defined by the unit ball

K = U ( p N ) = { x : x X 1 } .
Given an encoder/decoder pair ( Φ , Δ ) , the worst case error on a set K for that pair will be given by
E ( K , Φ , Δ ) X = sup x K x Δ ( Φ x ) X .
Finally, using min-max principles we will define the minimum error over all encoder/decoder pairs for a signal class and for a fixed number of measurements n to be
E n ( K ) X = inf ( Φ , Δ ) : Φ is n × N E ( K , Φ , Δ ) X .
This measure E n ( K ) X is the best we could do while measuring distortion on the topology of X , using n linear measurements, and using arbitrary decoding.

We will see that these questions are actually equivalent to a classical “ n -width” problem. n -widths have seen a great deal of work over the years by a variety of mathematicians: Kolmogorov, Tikhomirov, Kashin, Gluskin, etc. There are many different flavors of n -widths, but we will study the Gelfand n -width (the least intuitive of the widths).

Gelfand n-width

Let K X be compact. Given n , the Gelfand width (also called the dual width) is given by

d n ( K ) X : = inf Y : codim ( Y ) = n sup { x X : x K Y } .
where by codimension ( Y ) = n we mean that Y has dimension dim ( X ) n .

In other words, we are looking for the subspace Y that slices through the set K so that the norms of the projected signals are as small as possible. We can now state the following theorem about n-widths:

Provided that K has the properties (1) K = - K and (2) K + K = C K , then

d n ( K ) X E n ( K ) X C d n ( K ) X
where C is the same constant listed in property (2).
Clarifying notation: C K = { C x : x K } and K + K = { x 1 + x 2 : x 1 , x 2 K } .

We start with the left-hand inequality. We want to take any encoder/decoder pair and use that to construct a Y . So let Φ , Δ be an encoder/decoder. Then simply let Y = N ( Φ ) . Now consider an η K Y and note that Φ ( η ) = 0 since η Y . Let z = Δ ( 0 ) be the decoding of 0 (practically speaking, z should be zero itself, but we avoid that assumption in this proof). Then

η X max ( η - z X , η + z X ) = max ( η - Δ Φ ( η ) X , - η - Δ Φ ( η ) X ) sup x K x - Δ Φ ( x ) X
where we first employ the triangle inequality, then the fact that multiplying by - 1 does not change the norm, then the fact that K = - K . So then
sup η K Y η X sup x K x - Δ Φ ( x ) X .
Taking the infimum over all Φ , Δ , it follows that
d n ( K ) X E n ( K ) X .
Since Φ is n × N , then dim ( N ( Φ ) ) N - n .

Now we prove the right-hand inequality. Assume we have a good Y . Suppose Y has codimension N - n . Then Y (the orthogonal complement of Y in R N ) has dimension n . Let v 1 , v 2 , , v n R N be a basis for Y . Let Φ be the n × N matrix obtained by stacking the rows v 1 , v 2 , , v n . Then N ( Φ ) = Y . Define Δ ( y ) = any element of K F ( y ) if there is one (otherwise let Δ ( y ) be anything in F ( y ) ). Now look at the performance x - Δ Φ ( x ) X for some x K . Both x and Δ Φ ( x ) = : x ' are elements of K , so x - x ' is in N ( Φ ) and in C K . Therefore x - x ' C K N ( Φ ) . Thus,

x - x ' C X sup z Y K z X ,
and so for any x K ,
x - Δ Φ ( x ) X C sup z Y K z X .
Taking the infimum over all Y , we get that E n ( K ) X C d n ( K ) X .

From the proof of this theorem, we see that there is a matching between the matrices Φ and the spaces Y (via the nullspace of Φ ).

An important result is that d n ( U ( q N ) ) p N is known for all p , q except p = 1 , q = . A precise statement of these widths can be found in the book . A particularly important case is

C 1 log ( N / n ) n d n ( U ( 1 N ) ) 2 N C 2 log ( N / n ) n
for N > 2 n . This result was first proved by Kashin with a worselogarithm in the upper inequality and later brought to the present form by Gluskin. This result solves several important problems infunctional analysis and approximation.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Pdf generation problem modules. OpenStax CNX. Sep 23, 2008 Download for free at http://cnx.org/content/col10514/1.4
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Pdf generation problem modules' conversation and receive update notifications?

Ask