<< Chapter < Page Chapter >> Page >

A distribution function determines the probability mass in each semiinfinite interval ( - , t ] . According to the discussion referred to above, this determines uniquely the induced distribution.

The distribution function F X for a simple random variable is easily visualized. The distribution consists of point mass p i at each point t i in the range. To the left of the smallest value in the range, F X ( t ) = 0 ; as t increases to the smallest value t 1 , F X ( t ) remains constant at zero until it jumps by the amount p 1 . . F X ( t ) remains constant at p 1 until t increases to t 2 , where it jumps by an amount p 2 to the value p 1 + p 2 . This continues until the value of F X ( t ) reaches 1 at the largest value t n . The graph of F X is thus a step function, continuous from the right, with a jump in the amount p i at the corresponding point t i in the range. A similar situation exists for a discrete-valued random variable which may take on an infinity of values (e.g., the geometric distributionor the Poisson distribution considered below). In this case, there is always some probability at points to the right of any t i , but this must become vanishingly small as t increases, since the total probability mass is one.

The procedure ddbn may be used to plot the distributon function for a simple random variable from a matrix X of values and a corresponding matrix P X of probabilities.

Graph of F X For a simple random variable

>>c = [10 18 10 3]; % Distribution for X in Example 6.5.1>>pm = minprob(0.1*[6 3 5]);>>canonic Enter row vector of coefficients cEnter row vector of minterm probabilities pm Use row matrices X and PX for calculationsCall for XDBN to view the distribution>>ddbn % Circles show values at jumps Enter row matrix of VALUES XEnter row matrix of PROBABILITIES PX % Printing details See [link]
Got questions? Get instant answers now!
A graph of Distributive Function F. The range of u=F(t) is represented on the y-axis (0.1-1.1), while the t value is represented on the x-axis (0-50). The points that are plotted on the graph ascend to the right. A graph of Distributive Function F. The range of u=F(t) is represented on the y-axis (0.1-1.1), while the t value is represented on the x-axis (0-50). The points that are plotted on the graph ascend to the right.
Distribution function for [link]

Description of some common discrete distributions

We make repeated use of a number of common distributions which are used in many practical situations. This collection includes several distributions which are studied in the chapter "Random Variables and Probabilities" .

  1. Indicator function . X = I E P ( X = 1 ) = P ( E ) = p P ( X = 0 ) = q = 1 - p . The distribution function has a jump in the amount q at t = 0 and an additional jump of p to the value 1 at t = 1 .
  2. Simple random variable X = i = 1 n t i I A i (canonical form)
    P ( X = t i ) = P ( A i ) = p 1
    The distribution function is a step function, continuous from the right, with jump of p i at t = t i (See [link] for [link] )
  3. Binomial ( n , p ) . This random variable appears as the number of successes in a sequence of n Bernoulli trials with probability p of success. In its simplest form
    X = i = 1 n I E i with { E i : 1 i n } independent
    P ( E i ) = p P ( X = k ) = C ( n , k ) p k q n - k
    As pointed out in the study of Bernoulli sequences in the unit on Composite Trials, two m-functions ibinom and cbinom are available for computing the individual and cumulative binomial probabilities.
  4. Geometric ( p ) There are two related distributions, both arising in the study of continuing Bernoulli sequences. The first counts the number of failures before the first success. This is sometimes called the “waiting time.” The event { X = k } consists of a sequence of k failures, then a success. Thus
    P ( X = k ) = q k p , 0 k
    The second designates the component trial on which the first success occurs. The event { Y = k } consists of k - 1 failures, then a success on the k th component trial. We have
    P ( Y = k ) = q k - 1 p , 1 k
    We say X has the geometric distribution with parameter ( p ) , which we often designate by X geometric ( p ) . Now Y = X + 1 or Y - 1 = X . For this reason, it is customary to refer to the distribution for the number of the trial for the first successby saying Y - 1 geometric ( p ) . The probability of k or more failures before the first success is P ( X k ) = q k . Also
    P ( X n + k | X n ) = P ( X n + k ) P ( X n ) = q n + k / q n = q k = P ( X k )
    This suggests that a Bernoulli sequence essentially "starts over" on each trial. If it has failed n times, the probability of failing an additional k or more times before the next success is the same as the initial probability of failing k or more times before the first success.

    The geometric distribution

    A statistician is taking a random sample from a population in which two percent of the members own a BMW automobile. She takes a sample of size 100. What is the probabilityof finding no BMW owners in the sample?

    Solution

    The sampling process may be viewed as a sequence of Bernoulli trials with probability p = 0 . 02 of success. The probability of 100 or more failures before the first success is 0 . 98 100 = 0 . 1326 or about 1/7.5.

    Got questions? Get instant answers now!
  5. Negative binomial ( m , p ) . X is the number of failures before the m th success. It is generally more convenient to work with Y = X + m , the number of the trial on which the m th success occurs. An examination of the possible patterns and elementary combinatorics show that
    P ( Y = k ) = C ( k - 1 , m - 1 ) p m q k - m , m k
    There are m - 1 successes in the first k - 1 trials, then a success. Each combination has probability p m q k - m . We have an m-function nbinom to calculate these probabilities.

    A game of chance

    A player throws a single six-sided die repeatedly. He scores if he throws a 1 or a 6. What is the probability he scores five times in ten or fewer throws?

    >> p = sum(nbinom(5,1/3,5:10)) p  =  0.2131

    An alternate solution is possible with the use of the binomial distribution . The m th success comes not later than the k th trial iff the number of successes in k trials is greater than or equal to m .

    >> P = cbinom(10,1/3,5) P  =  0.2131
    Got questions? Get instant answers now!
  6. Poisson ( μ ) . This distribution is assumed in a wide variety of applications.It appears as a counting variable for items arriving with exponential interarrival times (see the relationship to the gamma distribution below). For large n and small p (which may not be a value found in a table), the binomial distribution is approximately Poisson ( n p ) . Use of the generating function (see Transform Methods) shows the sum ofindependent Poisson random variables is Poisson. The Poisson distribution is integer valued, with
    P ( X = k ) = e - μ μ k k ! 0 k
    Although Poisson probabilities are usually easier to calculate with scientific calculators than binomial probabilities, the use of tables is often quite helpful. As in thecase of the binomial distribution, we have two m-functions for calculating Poisson probabilities. These have advantages of speed and parameter range similar to those for ibinomand cbinom.
    • P ( X = k ) is calculated by P = ipoisson(mu,k) , where k is a row or column vector of integers and the result P is a row matrix of the probabilities.
    • P ( X k ) is calculated by P = cpoisson(mu,k) , where k is a row or column vector of integers and the result P is a row matrix of the probabilities.

    Poisson counting random variable

    The number of messages arriving in a one minute period at a communications network junction is a random variable N Poisson (130). What is the probability the number of arrivals is greater than equal to 110, 120, 130, 140, 150, 160 ?

    >> p = cpoisson(130,110:10:160) p  =  0.9666  0.8209  0.5117  0.2011  0.0461  0.0060

    The descriptions of these distributions, along with a number of other facts, are summarized in the table DATA ON SOME COMMON DISTRIBUTIONS in Appendix C .

    Got questions? Get instant answers now!

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Applied probability. OpenStax CNX. Aug 31, 2009 Download for free at http://cnx.org/content/col10708/1.6
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Applied probability' conversation and receive update notifications?

Ask