<< Chapter < Page Chapter >> Page >

In [link] , Donoho and Johnstone showed that, when using the universal threshold, the following oracle inequality prevails

R ( d ^ t , d ) ( 2 log n + 1 ) σ ϵ 2 n + R ideal .

However, this inequality is not optimal. Donoho and Johnstone looked for the optimal threshold t * ( n ) which leads to the smallest possible constant Λ n * in place of 2 log n + 1 . Such a threshold does not exist in closed form, but can be approximated numerically. For small sample size, it is sensibly smaller than the universal threshold.

Sureshrink procedure

Given the expression [link] for the l 2 -risk, it is natural to look for a threshold that minimizes an estimation of this risk.

By minimizing Stein's unbiased estimate of the risk [link] and using a soft thresholding scheme, the resulting estimator, called `SureShrink', is adaptive over a wide range of function spaces including Hölder, Sobolev, and Besov spaces, see "Adaptivity of wavelet estimator" . That is, without any a priori knowledge on the type or amount of regularity of the function, the SURE procedure nevertheless achieves the optimal rate of convergence that one could attain by knowing the regularity of the function.

Other thresholding procedures

We mention some of the other thresholding or shrinkage procedures proposed in the literature.

Instead of considering each coefficient individually, Cai et al. [link] , [link] consider blocks of empirical wavelet coefficients in order to make simultaneous shrinkagedecisions about all coefficients within a block.

Another fruitful idea is to use the Bayesian framework. There a prior distribution is imposed on the wavelet coefficients d j k . This prior model is designed to capture the sparseness of the wavelet expansion. Next, the function is estimated by applying some Bayes rules on the resulting posteriordistribution of the wavelet coefficients, see for example [link] , [link] , [link] , [link] .

Antoniadis and Fan [link] treat the problem of selecting the wavelet coefficients as a penalized least squares issue. Let W be the matrix of an orthogonal wavelet transform and Y : = { Y i } i = 1 n . The detail coefficients d : = { d j k } which minimize

W Y - d l 2 2 + j , k p λ ( | d j k | )

are used to estimate the true wavelet coefficients. In equation [link] , p λ ( · ) is a penalty function which depends on the regularization parameter λ . The authors provide a general framework, wheredifferent penalty functions correspond to different type of thresholding procedures (like, e.g., the soft- and hard- thresholding)and obtain oracle inequalities for a large class of penalty functions.

Other methods include threshold selection by hypothesis testing [link] , cross-validation [link] , or generalizedcross-validation [link] , [link] , which is used to estimated the l 2 -risk of the empirical detail coefficients.

Linear versus nonlinear wavelet estimator

In order to differenciate the behaviours of a linear and of a nonlinear wavelet estimator, we consider the Sobolev class W q s ( C ) defined as

W q s ( C ) = { f : f q q + d s d x s f ( x ) q q C 2 } ,

and that we denote V in short. Assume we know that m , the function to be estimated, belongs to V . In the next section, we will release this assumption. The L p - risk of an arbitrary estimator T n based on the sample data is defined as E T n - m p p , 1 p < , whereas the L p - minimax risk is given by

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, An introduction to wavelet analysis. OpenStax CNX. Sep 14, 2009 Download for free at http://cnx.org/content/col10566/1.3
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'An introduction to wavelet analysis' conversation and receive update notifications?

Ask