<< Chapter < Page Chapter >> Page >

Properties and computation

The properties in the table in Appendix E constitute a powerful and convenient resource for the use of mathematical expectation. These are properties of the abstract Lebesgueintegral, expressed in the notation for mathematical expectation.

E [ g ( X ) ] = g ( X ) d P

In the development of additional properties, the four basic properties: (E1) Expectation of indicator functions, (E2) Linearity, (E3) Positivity; monotonicity, and (E4a) Monotone convergence play a foundational role. We utilize the properties in the table, as needed, often referring to them by the numbers assigned in the table.

In this section, we include a number of examples which illustrate the use of various properties. Some are theoretical examples, deriving additional properties or displaying the basis and structureof some in the table. Others apply these properties to facilitate computation

Probability as expectation

Probability may be expressed entirely in terms of expectation.

  • By properties (E1) and positivity (E3a) , P ( A ) = E [ I A ] 0 .
  • As a special case of (E1) , we have P ( Ω ) = E [ I Ω ] = 1
  • By the countable sums property (E8) ,
    A = i A i implies P ( A ) = E [ I A ] = E [ i I A i ] = i E [ I A i ] = i P ( A i )

Thus, the three defining properties for a probability measure are satisfied.

Got questions? Get instant answers now!

Remark . There are treatments of probability which characterize mathematical expectation with properties (E0) through (E4a) , then define P ( A ) = E [ I A ] . Although such a development is quite feasible, it has not been widely adopted.

An indicator function pattern

Suppose X is a real random variable and E = X - 1 ( M ) = { ω : X ( ω ) M } . Then

I E = I M ( X )

To see this, note that X ( ω ) M iff ω E , so that I E ( ω ) = 1 iff I M ( X ( ω ) ) = 1 .

Similarly, if E = X - 1 ( M ) Y - 1 ( N ) , then I E = I M ( X ) I N ( Y ) . We thus have, by (E1) ,

P ( X M ) = E [ I M ( X ) ] and P ( X M , Y N ) = E [ I M ( X ) I N ( Y ) ]
Got questions? Get instant answers now!

Alternate interpretation of the mean value

E [ ( X - c ) 2 ] is a minimum iff c = E [ X ] , in which case E ( X - E [ X ] ) 2 = E [ X 2 ] - E 2 [ X ]

INTERPRETATION. If we approximate the random variable X by a constant c , then for any ω the error of approximation is X ( ω ) - c . The probability weighted average of the square of the error (often called the mean squared error ) is E ( X - c ) 2 . This average squared error is smallest iff the approximating constant c is the mean value.

VERIFICATION

We expand ( X - c ) 2 and apply linearity to obtain

E ( X - c ) 2 = E [ X 2 - 2 c X + c 2 ] = E [ X 2 ] - 2 E [ X ] c + c 2

The last expression is a quadratic in c (since E [ X 2 ] and E [ X ] are constants). The usual calculus treatment shows the expression has a minimum for c = E [ X ] . Substitution of this value for c shows the expression reduces to E [ X 2 ] - E 2 [ X ] .

Got questions? Get instant answers now!

A number of inequalities are listed among the properties in the table. The basis for these inequalities is usually some standard analytical inequality on random variables to which themonotonicity property is applied. We illustrate with a derivation of the important Jensen's inequality.

Jensen's inequality

If X is a real random variable and g is a convex function on an interval I which includes the range of X , then

g ( E [ X ] ) E [ g ( X ) ]

VERIFICATION

The function g is convex on I iff for each t 0 I = [ a , b ] there is a number λ ( t 0 ) such that

g ( t ) g ( t 0 ) + λ ( t 0 ) ( t - t 0 )

This means there is a line through ( t 0 , g ( t 0 ) ) such that the graph of g lies on or above it. If a X b , then by monotonicity E [ a ] = a E [ X ] E [ b ] = b (this is the mean value property (E11) ). We may choose t 0 = E [ X ] I . If we designate the constant λ ( E [ X ] ) by c , we have

g ( X ) g ( E [ X ] ) + c ( X - E [ X ] )

Recalling that E [ X ] is a constant, we take expectation of both sides, using linearity and monotonicity, to get

E [ g ( X ) ] g ( E [ X ] ) + c ( E [ X ] - E [ X ] ) = g ( E [ X ] )
Got questions? Get instant answers now!

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Applied probability. OpenStax CNX. Aug 31, 2009 Download for free at http://cnx.org/content/col10708/1.6
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Applied probability' conversation and receive update notifications?

Ask