<< Chapter < Page Chapter >> Page >
This module introduces estimation theory and its terminology, including bias, consistency, and efficiency.

In searching for methods of extracting information from noisy observations, this chapter describes estimation theory , which has the goal of extracting from noise-corrupted observations the values of disturbanceparameters (noise variance, for example), signal parameters (amplitude or propagation direction), or signalwaveforms . Estimation theory assumes that the observations contain an information-bearing quantity, therebytacitly assuming that detection-based preprocessing has been performed (in other words, do I have something in theobservations worth estimating?). Conversely, detection theory often requires estimation of unknown parameters: Signal presenceis assumed, parameter estimates are incorporated into the detection statistic, and consistency of observations andassumptions tested. Consequently, detection and estimation theory form a symbiotic relationship, each requiring the otherto yield high-quality signal processing algorithms.

Despite a wide variety of error criteria and problem frameworks, the optimal detector is characterized by a single result: thelikelihood ratio test. Surprisingly, optimal detectors thus derived are usually easy to implement, not often requiringsimplification to obtain a feasible realization in hardware or software. In contrast to detection theory, no fundamentalresult in estimation theory exists to be summoned to attack the problem at hand. The choice of error criterion and itsoptimization heavily influences the form of the estimation procedure. Because of the variety of criterion-dependentestimators, arguments frequently rage about which of several optimal estimators is "better." Each procedure is optimum forits assumed error criterion; thus, the argument becomes which error criterion best describes some intuitive notion of quality.When more ad hoc, noncriterion-based procedures

This governmentese phrase concisely means guessing.
are used, we cannot assess the quality of the resulting estimator relative to the best achievable. As shown later , bounds on the estimation error do exist, but their tightness and applicability to a givensituation are always issues in assessing estimator quality. At best, estimation theory is less structured than detectiontheory. Detection is science, estimation art. Inventiveness coupled with an understanding of the problem (what types oferrors are critically important, for example) are key elements to deciding which estimation procedure "fits" a given problemwell.

Terminology in estimation theory

More so than detection theory, estimation theory relies on jargon to characterize the properties of estimators. Withoutknowing any estimation technique, let's use parameter estimation as our discussion prototype. The parameterestimation problem is to determine from a set of L observations, represented by the L -dimensional vector r , the values of parameters denoted by the vector . We write the estimate of this parameter vector as r , where the "hat" denotes the estimate, and the functional dependence on r explicitly denotes the dependence of the estimate on the observations. Thisdependence is always present

Estimating the value of a parameter given no data may be an interesting problem inclairvoyance, but not in estimation theory.
, but we frequently denote the estimate compactly as . Because of the probabilistic nature of the problems considered in this chapter, a parameter estimate isitself a random vector, having its own statistical characteristics. The estimation error r equals the estimate minus the actual parameter value: r r . It too is a random quantity and is often used in the criterion function. For example, the mean-squared error is given by ; the minimum mean-squared error estimate would minimize this quantity. The mean-squared error matrix is ; on the main diagonal, its entries are the mean-squared estimation errors for each component of theparameter vector, whereas the off-diagonal terms express the correlation between the errors. The mean-squared estimation error equals the trace of the mean-squared error matrix tr .

Bias

An estimate is said to be unbiased if the expected value of the estimate equals the true value of theparameter: . Otherwise, the estimate is said to be biased : . The bias b is usually considered to be additive, so that b . When we have a biased estimate, the bias usually depends on the number of observations L . An estimate is said to be asymptotically unbiased if the bias tends to zero for large L : L b 0 . An estimate's variance equals the mean-squared estimation error only if the estimate is unbiased.

An unbiased estimate has a probability distribution where the mean equals the actual value of the parameter. Shouldthe lack of bias be considered a desirable property? If many unbiased estimates are computed from statisticallyindependent sets of observations having the same parameter value, the average of these estimates will be close to thisvalue. This property does not mean that the estimate has less error than a biased one; thereexist biased estimates whose mean-squared errors are smaller than unbiased ones. In such cases, the biased estimate isusually asymptotically unbiased. Lack of bias is good, but that is just one aspect of how we evaluate estimators.

Consistency

We term an estimate consistent if the mean-squared estimation error tends to zero as the number ofobservations becomes large: L 0 . Thus, a consistent estimate must be at least asymptotically unbiased. Unbiased estimates do exist whoseerrors never diminish as more data are collected: Their variances remain nonzero no matter how much data areavailable. Inconsistent estimates may provide reasonable estimates when the amount of data is limited, but have thecounterintuitive property that the quality of the estimate does not improve as the number of observations increases.Although appropriate in the proper circumstances (smaller mean-squared error than a consistent estimate over apertinent range of values of L , consistent estimates are usually favored in practice.

Efficiency

As estimators can be derived in a variety of ways, their error characteristics must always be analyzed and compared.In practice, many problems and the estimators derived for them are sufficiently complicated to render analytic studiesof the errors difficult, if not impossible. Instead, numerical simulation and comparison with lower bounds on theestimation error are frequently used instead to assess the estimator performance. An efficient estimate has a mean-squared error that equals a particular lower bound: the Cramr-Rao bound . If an efficient estimate exists (the Cramr-Rao bound is the greatest lower bound), it is optimum in the mean-squaredsense: No other estimate has a smaller mean-squared error (see Maximum Likelihood Estimators for details).

For many problems no efficient estimate exists. In such cases, the Cramr-Rao bound remains a lower bound, but its value is smaller than that achievable by anyestimator. How much smaller is usually not known. However, practitioners frequently use the Cramr-Rao bound in comparisons with numerical error calculations. Anotherissue is the choice of mean-squared error as the estimation criterion; it may not suffice to pointedly assess estimatorperformance in a particular problem. Nevertheless, every problem is usually subjected to a Cramr-Rao bound computation and the existence of an efficient estimate considered.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Signal and information processing for sonar. OpenStax CNX. Dec 04, 2007 Download for free at http://cnx.org/content/col10422/1.5
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Signal and information processing for sonar' conversation and receive update notifications?

Ask