Jump to content

Geometric distribution

From Wikipedia, the free encyclopedia
Geometric
Probability mass function
Cumulative distribution function
Parameters success probability (real) success probability (real)
Support k trials where k failures where
PMF
CDF for ,
for
for ,
for
Mean
Median


(not unique if is an integer)


(not unique if is an integer)
Mode
Variance
Skewness
Excess kurtosis
Entropy
MGF
for

for
CF
PGF

In probability theory and statistics, the geometric distribution is either one of two discrete probability distributions:

  • The probability distribution of the number of Bernoulli trials needed to get one success, supported on the set ;
  • The probability distribution of the number of failures before the first success, supported on the set .

Which of these is called the geometric distribution is a matter of convention and convenience.

These two different geometric distributions should not be confused with each other. Often, the name shifted geometric distribution is adopted for the former one (distribution of ); however, to avoid ambiguity, it is considered wise to indicate which is intended, by mentioning the support explicitly.

The geometric distribution gives the probability that the first occurrence of success requires independent trials, each with success probability . If the probability of success on each trial is , then the probability that the -th trial is the first success is

for

The above form of the geometric distribution is used for modeling the number of trials up to and including the first success. By contrast, the following form of the geometric distribution is used for modeling the number of failures until the first success:

for

In either case, the sequence of probabilities is a geometric sequence.

The geometric distribution is denoted by Geo(p) where .[1]

Definition[edit]

The geometric distribution is the discrete probability distribution that describes when the first success in an infinite sequence of independent and identically distributed Bernoulli trials occurs. Its probability mass function is

where is the number of trials and is the probability of success in each trial.[2]

Alternatively, some texts define the distribution where altering the probability mass function into:[2]

An example of a geometric distribution comes from rolling a six-sided die until a "1" appears. Each roll is independent with a chance of success. The number of rolls needed follows a geometric distribution with .

Properties[edit]

Moments and cumulants[edit]

The expected value for the number of independent trials to get the first success, and the variance of a geometrically distributed random variable X is:

Similarly, the expected value and variance of the geometrically distributed random variable Y = X - 1 (See definition of distribution ) is:

Proof[edit]

Expected value of X[edit]

Consider the expected value of X as above, i.e. the average number of trials until a success. On the first trial we either succeed with probability , or we fail with probability . If we fail the remaining mean number of trials until a success is identical to the original mean. This follows from the fact that all trials are independent. From this we get the formula:

which if solved for gives:

Expected value of Y[edit]

That the expected value of Y as above is (1 − p)/p can be trivially seen from which follows from the linearity of expectation, or can be shown in the following way:

The interchange of summation and differentiation is justified by the fact that convergent power series converge uniformly on compact subsets of the set of points where they converge.

Let μ = (1 − p)/p be the expected value of Y. Then the cumulants of the probability distribution of Y satisfy the recursion

Expected value examples[edit]

E3) A patient is waiting for a suitable matching kidney donor for a transplant. If the probability that a randomly selected donor is a suitable match is p = 0.1, what is the expected number of donors who will be tested before a matching donor is found?

With p = 0.1, the mean number of failures before the first success is E(Y) = (1 − p)/p =(1 − 0.1)/0.1 = 9.

For the alternative formulation, where X is the number of trials up to and including the first success, the expected value is E(X) = 1/p = 1/0.1 = 10.

For example 1 above, with p = 0.6, the mean number of failures before the first success is E(Y) = (1 − p)/p = (1 − 0.6)/0.6 = 0.67.

Higher-order moments[edit]

The moments for the number of failures before the first success are given by

where is the polylogarithm function.

General properties[edit]

The geometric distribution supported on {0, 1, 2, 3, ... } is the only memoryless discrete distribution. Note that the geometric distribution supported on {1, 2, ... } is not memoryless.
  • Among all discrete probability distributions supported on {1, 2, 3, ... } with given expected value μ, the geometric distribution X with parameter p = 1/μ is the one with the largest entropy.[3]
  • The geometric distribution of the number Y of failures before the first success is infinitely divisible, i.e., for any positive integer n, there exist independent identically distributed random variables Y1, ..., Yn whose sum has the same distribution that Y has. These will not be geometrically distributed unless n = 1; they follow a negative binomial distribution.
  • The decimal digits of the geometrically distributed random variable Y are a sequence of independent (and not identically distributed) random variables.[citation needed] For example, the hundreds digit D has this probability distribution:
where q = 1 − p, and similarly for the other digits, and, more generally, similarly for numeral systems with other bases than 10. When the base is 2, this shows that a geometrically distributed random variable can be written as a sum of independent random variables whose probability distributions are indecomposable.

Related distributions[edit]

  • The geometric distribution Y is a special case of the negative binomial distribution, with r = 1. More generally, if Y1, ..., Yr are independent geometrically distributed variables with parameter p, then the sum
follows a negative binomial distribution with parameters r and p.[5]
  • The geometric distribution is a special case of discrete compound Poisson distribution.
  • If Y1, ..., Yr are independent geometrically distributed variables (with possibly different success parameters pm), then their minimum
is also geometrically distributed, with parameter [6]
  • Suppose 0 < r < 1, and for k = 1, 2, 3, ... the random variable Xk has a Poisson distribution with expected value rk/k. Then
has a geometric distribution taking values in the set {0, 1, 2, ...}, with expected value r/(1 − r).[citation needed]
  • The exponential distribution is the continuous analogue of the geometric distribution. If X is an exponentially distributed random variable with parameter λ, then
where is the floor (or greatest integer) function, is a geometrically distributed random variable with parameter p = 1 − eλ (thus λ = −ln(1 − p)[7]) and taking values in the set {0, 1, 2, ...}. This can be used to generate geometrically distributed pseudorandom numbers by first generating exponentially distributed pseudorandom numbers from a uniform pseudorandom number generator: then is geometrically distributed with parameter , if is uniformly distributed in [0,1].
  • If p = 1/n and X is geometrically distributed with parameter p, then the distribution of X/n approaches an exponential distribution with expected value 1 as n → ∞, since

More generally, if p = λ/n, where λ is a parameter, then as n→ ∞ the distribution of X/n approaches an exponential distribution with rate λ:

therefore the distribution function of X/n converges to , which is that of an exponential random variable.

Statistical inference[edit]

Parameter estimation[edit]

For both variants of the geometric distribution, the parameter p can be estimated by equating the expected value with the sample mean. This is the method of moments, which in this case happens to yield maximum likelihood estimates of p.[8][9]

Specifically, for the first variant let k = k1, ..., kn be a sample where ki ≥ 1 for i = 1, ..., n. Then p can be estimated as

In Bayesian inference, the Beta distribution is the conjugate prior distribution for the parameter p. If this parameter is given a Beta(αβ) prior, then the posterior distribution is

The posterior mean E[p] approaches the maximum likelihood estimate as α and β approach zero.

In the alternative case, let k1, ..., kn be a sample where ki ≥ 0 for i = 1, ..., n. Then p can be estimated as

The posterior distribution of p given a Beta(αβ) prior is[10]

Again the posterior mean E[p] approaches the maximum likelihood estimate as α and β approach zero.

For either estimate of using Maximum Likelihood, the bias is equal to

which yields the bias-corrected maximum likelihood estimator

Computational methods[edit]

In the programming language R, the function dgeom(k, prob) calculates the probability of k failures before a success with a success probability prob for each trial.

In Microsoft Excel, the function NEGBINOMDIST(number_f, number_s, probability_s) can be used to calculate the number of failures, number_f, before a number of successes, number_s, with a success probability, probability_s, for each trial. Setting number_s to 1, gives the geometric distribution.[11]

See also[edit]

References[edit]

  1. ^ a b A modern introduction to probability and statistics : understanding why and how. Dekking, Michel, 1946-. London: Springer. 2005. pp. 48–50, 61–62, 152. ISBN 9781852338961. OCLC 262680588.{{cite book}}: CS1 maint: others (link)
  2. ^ a b Nagel, Werner; Steyer, Rolf (2017-04-04). Probability and Conditional Expectation: Fundamentals for the Empirical Sciences. Wiley Series in Probability and Statistics (1st ed.). Wiley. pp. 260–261. doi:10.1002/9781119243496. ISBN 978-1-119-24352-6.
  3. ^ Park, Sung Y.; Bera, Anil K. (June 2009). "Maximum entropy autoregressive conditional heteroskedasticity model". Journal of Econometrics. 150 (2): 219–230. doi:10.1016/j.jeconom.2008.12.014.
  4. ^ Gallager, R.; van Voorhis, D. (March 1975). "Optimal source codes for geometrically distributed integer alphabets (Corresp.)". IEEE Transactions on Information Theory. 21 (2): 228–230. doi:10.1109/TIT.1975.1055357. ISSN 0018-9448.
  5. ^ Pitman, Jim. Probability (1993 edition). Springer Publishers. pp 372.
  6. ^ Ciardo, Gianfranco; Leemis, Lawrence M.; Nicol, David (1 June 1995). "On the minimum of independent geometrically distributed random variables". Statistics & Probability Letters. 23 (4): 313–326. doi:10.1016/0167-7152(94)00130-Z. hdl:2060/19940028569. S2CID 1505801.
  7. ^ "Wolfram-Alpha: Computational Knowledge Engine". www.wolframalpha.com.
  8. ^ casella, george; berger, roger l (2002). statistical inference (2nd ed.). pp. 312–315. ISBN 0-534-24312-6.
  9. ^ "MLE Examples: Exponential and Geometric Distributions Old Kiwi - Rhea". www.projectrhea.org. Retrieved 2019-11-17.
  10. ^ "3. Conjugate families of distributions" (PDF). Archived (PDF) from the original on 2010-04-08.
  11. ^ "3.5 Geometric Probability Distribution using Excel Spreadsheet". Statistics LibreTexts. 2021-07-24. Retrieved 2023-10-20.

External links[edit]