<< Chapter < Page Chapter >> Page >

But why do the two algorithms converge to different places? The facile answer is that they are different becausethey minimize different performance functions. Indeed, the error surfaces in [link] show minima in different locations. The convergent value of a 0 . 38 for J N ( a ) is explicable because 0 . 38 2 0 . 15 = s 2 . The convergent value of a = 0 . 22 for J L S ( a ) is calculated in closed form in Exercise  [link] , and this value does a good job minimizing its cost,but it has not solved the problem of making a 2 close to s 2 . Rather, J L S ( a ) calculates a smaller gain that makes avg { s 2 } s 2 . The minima are different. The moral is this:Be wary of your performance functions—they may do what you ask.

The error surface for the AGC objective functions Equation 30 and Equation 34 each have two minima. As long as a can be initialized with the correct (positive) sign, there is little danger of converging to the wrong minimum.
The error surface for the AGC objective functions [link] and [link] each have two minima. As long as a can be initialized with the correct (positive) sign, there is little danger of converging to the wrongminimum.

Use agcgrad.m to investigate the AGC algorithm.

  1. What range of stepsize mu works? Can the stepsize be too small?Can the stepsize be too large?
  2. How does the stepsize mu effect the convergence rate?
  3. How does the variance of the input effect the convergent value of a ?
  4. What range of averages lenavg works? Can lenavg be too small? Can lenavg be too large?
  5. How does lenavg effect the convergence rate?

Show that the value of a that achieves the minimum of J L S ( a ) can be expressed as

± s 2 k r k 2 k r k 4 .

Is there a way to use this (closed form) solution to replace the iteration [link] ?

Consider the alternative objective function J ( a ) = 1 2 a 2 ( 1 2 s 2 [ k ] 3 - s 2 ) . Calculate the derivative and implement avariation of the AGC algorithm that minimizes this objective. How does this version compare to the algorithms [link] and [link] ? Draw the error surface for this algorithm. Which version is preferable?

Try initializing the estimate a(1)=-2 in agcgrad.m . Which minimum does the algorithm find? What happens tothe data record?

Create your own objective function J ( a ) for the AGC problem. Calculate the derivative and implement avariation of the AGC algorithm that minimizes this objective. How does this version compare to the algorithms [link] and [link] ? Draw the error surface for your algorithm. Which version do you prefer?

Investigate how the error surface depends on the input signal. Replace randn with rand in agcerrorsurf.m and draw the error surfaces for both J N ( a ) and J L S ( a ) .

Using an agc to combat fading

One of the impairments encountered in transmission systems is the degradation due to fading, when the strengthof the received signal changes in response to changes in the transmission path. (Recall the discussion in [link] .) This section shows how an AGC can be used to counteractthe fading, assuming the rate of the fading is slow, and provided the signal does not disappear completely.

Suppose that the input consists of a random sequence undulating slowly up and down in magnitude, as in the topplot of [link] . The adaptive AGC compensates for the amplitude variations,growing small when the power of the input is large, and large when the power of the input is small. This is shown in themiddle graph. The resulting output is of roughly constant amplitude, as shown in the bottom plot of [link] .

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Software receiver design. OpenStax CNX. Aug 13, 2013 Download for free at http://cnx.org/content/col11510/1.3
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Software receiver design' conversation and receive update notifications?

Ask