(Translated by https://www.hiragana.jp/)
Relationships among probability distributions - Wikipedia

Relationships among probability distributions

In probability theory and statistics, there are several relationships among probability distributions. These relations can be categorized in the following groups:

  • One distribution is a special case of another with a broader parameter space
  • Transforms (function of a random variable);
  • Combinations (function of several variables);
  • Approximation (limit) relationships;
  • Compound relationships (useful for Bayesian inference);
  • Duality[clarification needed];
  • Conjugate priors.
Relationships among some of univariate probability distributions are illustrated with connected lines. dashed lines means approximate relationship. more info:[1]
Relationships between univariate probability distributions in ProbOnto.[2]

Special case of distribution parametrization

edit

Transform of a variable

edit

Multiple of a random variable

edit

Multiplying the variable by any positive real constant yields a scaling of the original distribution. Some are self-replicating, meaning that the scaling yields the same family of distributions, albeit with a different parameter: normal distribution, gamma distribution, Cauchy distribution, exponential distribution, Erlang distribution, Weibull distribution, logistic distribution, error distribution, power-law distribution, Rayleigh distribution.

Example:

  • If X is a gamma random variable with shape and rate parameters (αあるふぁ, βべーた), then Y = aX is a gamma random variable with parameters (αあるふぁ,βべーた/a).
  • If X is a gamma random variable with shape and scale parameters (k, θしーた), then Y = aX is a gamma random variable with parameters (k,aθしーた).

Linear function of a random variable

edit

The affine transform ax + b yields a relocation and scaling of the original distribution. The following are self-replicating: Normal distribution, Cauchy distribution, Logistic distribution, Error distribution, Power distribution, Rayleigh distribution.

Example:

  • If Z is a normal random variable with parameters (μみゅー = m, σしぐま2 = s2), then X = aZ + b is a normal random variable with parameters (μみゅー = am + b, σしぐま2 = a2s2).

Reciprocal of a random variable

edit

The reciprocal 1/X of a random variable X, is a member of the same family of distribution as X, in the following cases: Cauchy distribution, F distribution, log logistic distribution.

Examples:

  • If X is a Cauchy (μみゅー, σしぐま) random variable, then 1/X is a Cauchy (μみゅー/C, σしぐま/C) random variable where C = μみゅー2 + σしぐま2.
  • If X is an F(νにゅー1, νにゅー2) random variable then 1/X is an F(νにゅー2, νにゅー1) random variable.

Other cases

edit

Some distributions are invariant under a specific transformation.

Example:

  • If X is a beta (αあるふぁ, βべーた) random variable then (1 − X) is a beta (βべーた, αあるふぁ) random variable.
  • If X is a binomial (n, p) random variable then (nX) is a binomial (n, 1 − p) random variable.
  • If X has cumulative distribution function FX, then the inverse of the cumulative distribution F
    X
    (X) is a standard uniform (0,1) random variable
  • If X is a normal (μみゅー, σしぐま2) random variable then eX is a lognormal (μみゅー, σしぐま2) random variable.
Conversely, if X is a lognormal (μみゅー, σしぐま2) random variable then log X is a normal (μみゅー, σしぐま2) random variable.
  • If X is an exponential random variable with mean βべーた, then X1/γがんま is a Weibull (γがんま, βべーた) random variable.
  • The square of a standard normal random variable has a chi-squared distribution with one degree of freedom.
  • If X is a Student’s t random variable with νにゅー degree of freedom, then X2 is an F (1,νにゅー) random variable.
  • If X is a double exponential random variable with mean 0 and scale λらむだ, then |X| is an exponential random variable with mean λらむだ.
  • A geometric random variable is the floor of an exponential random variable.
  • A rectangular random variable is the floor of a uniform random variable.
  • A reciprocal random variable is the exponential of a uniform random variable.

Functions of several variables

edit

Sum of variables

edit

The distribution of the sum of independent random variables is the convolution of their distributions. Suppose   is the sum of   independent random variables   each with probability mass functions  . Then

 

If it has a distribution from the same family of distributions as the original variables, that family of distributions is said to be closed under convolution. Often (always?) these distributions are also stable distributions (see also Discrete-stable distribution).

Examples of such univariate distributions are: normal distributions, Poisson distributions, binomial distributions (with common success probability), negative binomial distributions (with common success probability), gamma distributions (with common rate parameter), chi-squared distributions, Cauchy distributions, hyperexponential distributions.

Examples:[3][4]

    • If X1 and X2 are Poisson random variables with means μみゅー1 and μみゅー2 respectively, then X1 + X2 is a Poisson random variable with mean μみゅー1 + μみゅー2.
    • The sum of gamma (αあるふぁi, βべーた) random variables has a gamma (Σしぐまαあるふぁi, βべーた) distribution.
    • If X1 is a Cauchy (μみゅー1, σしぐま1) random variable and X2 is a Cauchy (μみゅー2, σしぐま2), then X1 + X2 is a Cauchy (μみゅー1 + μみゅー2, σしぐま1 + σしぐま2) random variable.
    • If X1 and X2 are chi-squared random variables with νにゅー1 and νにゅー2 degrees of freedom respectively, then X1 + X2 is a chi-squared random variable with νにゅー1 + νにゅー2 degrees of freedom.
    • If X1 is a normal (μみゅー1, σしぐま2
      1
      ) random variable and X2 is a normal (μみゅー2, σしぐま2
      2
      ) random variable, then X1 + X2 is a normal (μみゅー1 + μみゅー2, σしぐま2
      1
      + σしぐま2
      2
      ) random variable.
    • The sum of N chi-squared (1) random variables has a chi-squared distribution with N degrees of freedom.

Other distributions are not closed under convolution, but their sum has a known distribution:

  • The sum of n Bernoulli (p) random variables is a binomial (n, p) random variable.
  • The sum of n geometric random variables with probability of success p is a negative binomial random variable with parameters n and p.
  • The sum of n exponential (βべーた) random variables is a gamma (n, βべーた) random variable. Since n is an integer, the gamma distribution is also a Erlang distribution.
  • The sum of the squares of N standard normal random variables has a chi-squared distribution with N degrees of freedom.

Product of variables

edit

The product of independent random variables X and Y may belong to the same family of distribution as X and Y: Bernoulli distribution and log-normal distribution.

Example:

  • If X1 and X2 are independent log-normal random variables with parameters (μみゅー1, σしぐま2
    1
    ) and (μみゅー2, σしぐま2
    2
    ) respectively, then X1 X2 is a log-normal random variable with parameters (μみゅー1 + μみゅー2, σしぐま2
    1
    + σしぐま2
    2
    ).

(See also Product distribution.)

Minimum and maximum of independent random variables

edit

For some distributions, the minimum value of several independent random variables is a member of the same family, with different parameters: Bernoulli distribution, Geometric distribution, Exponential distribution, Extreme value distribution, Pareto distribution, Rayleigh distribution, Weibull distribution.

Examples:

  • If X1 and X2 are independent geometric random variables with probability of success p1 and p2 respectively, then min(X1, X2) is a geometric random variable with probability of success p = p1 + p2p1 p2. The relationship is simpler if expressed in terms probability of failure: q = q1 q2.
  • If X1 and X2 are independent exponential random variables with rate μみゅー1 and μみゅー2 respectively, then min(X1, X2) is an exponential random variable with rate μみゅー = μみゅー1 + μみゅー2.

Similarly, distributions for which the maximum value of several independent random variables is a member of the same family of distribution include: Bernoulli distribution, Power law distribution.

Other

edit
  • If X and Y are independent standard normal random variables, X/Y is a Cauchy (0,1) random variable.
  • If X1 and X2 are independent chi-squared random variables with νにゅー1 and νにゅー2 degrees of freedom respectively, then (X1/νにゅー1)/(X2/νにゅー2) is an F(νにゅー1, νにゅー2) random variable.
  • If X is a standard normal random variable and U is an independent chi-squared random variable with νにゅー degrees of freedom, then   is a Student's t(νにゅー) random variable.
  • If X1 is a gamma (αあるふぁ1, 1) random variable and X2 is an independent gamma (αあるふぁ2, 1) random variable then X1/(X1 + X2) is a beta(αあるふぁ1, αあるふぁ2) random variable. More generally, if X1 is a gamma(αあるふぁ1, βべーた1) random variable and X2 is an independent gamma(αあるふぁ2, βべーた2) random variable then βべーた2 X1/(βべーた2 X1 + βべーた1 X2) is a beta(αあるふぁ1, αあるふぁ2) random variable.
  • If X and Y are independent exponential random variables with mean μみゅー, then X − Y is a double exponential random variable with mean 0 and scale μみゅー.
  • If Xi are independent Bernoulli random variables then their parity (XOR) is a Bernoulli variable described by the piling-up lemma.

(See also ratio distribution.)

Approximate (limit) relationships

edit

Approximate or limit relationship means

  • either that the combination of an infinite number of iid random variables tends to some distribution,
  • or that the limit when a parameter tends to some value approaches to a different distribution.

Combination of iid random variables:

  • Given certain conditions, the sum (hence the average) of a sufficiently large number of iid random variables, each with finite mean and variance, will be approximately normally distributed. This is the central limit theorem (CLT).

Special case of distribution parametrization:

  • X is a hypergeometric (m, N, n) random variable. If n and m are large compared to N, and p = m/N is not close to 0 or 1, then X approximately has a Binomial(n, p) distribution.
  • X is a beta-binomial random variable with parameters (n, αあるふぁ, βべーた). Let p = αあるふぁ/(αあるふぁ + βべーた) and suppose αあるふぁ + βべーた is large, then X approximately has a binomial(n, p) distribution.
  • If X is a binomial (n, p) random variable and if n is large and np is small then X approximately has a Poisson(np) distribution.
  • If X is a negative binomial random variable with r large, P near 1, and r(1 − P) = λらむだ, then X approximately has a Poisson distribution with mean λらむだ.

Consequences of the CLT:

  • If X is a Poisson random variable with large mean, then for integers j and k, P(jXk) approximately equals to P(j − 1/2 ≤ Yk + 1/2) where Y is a normal distribution with the same mean and variance as X.
  • If X is a binomial(n, p) random variable with large np and n(1 − p), then for integers j and k, P(jXk) approximately equals to P(j − 1/2 ≤ Yk + 1/2) where Y is a normal random variable with the same mean and variance as X, i.e. np and np(1 − p).
  • If X is a beta random variable with parameters αあるふぁ and βべーた equal and large, then X approximately has a normal distribution with the same mean and variance, i. e. mean αあるふぁ/(αあるふぁ + βべーた) and variance αあるふぁβべーた/((αあるふぁ + βべーた)2(αあるふぁ + βべーた + 1)).
  • If X is a gamma(αあるふぁ, βべーた) random variable and the shape parameter αあるふぁ is large relative to the scale parameter βべーた, then X approximately has a normal random variable with the same mean and variance.
  • If X is a Student's t random variable with a large number of degrees of freedom νにゅー then X approximately has a standard normal distribution.
  • If X is an F(νにゅー, ωおめが) random variable with ωおめが large, then νにゅーX is approximately distributed as a chi-squared random variable with νにゅー degrees of freedom.

Compound (or Bayesian) relationships

edit

When one or more parameter(s) of a distribution are random variables, the compound distribution is the marginal distribution of the variable.

Examples:

  • If X | N is a binomial (N,p) random variable, where parameter N is a random variable with negative-binomial (m, r) distribution, then X is distributed as a negative-binomial (m, r/(p + qr)).
  • If X | N is a binomial (N,p) random variable, where parameter N is a random variable with Poisson(μみゅー) distribution, then X is distributed as a Poisson (μみゅーp).
  • If X | μみゅー is a Poisson(μみゅー) random variable and parameter μみゅー is random variable with gamma(m, θしーた) distribution (where θしーた is the scale parameter), then X is distributed as a negative-binomial (m, θしーた/(1 + θしーた)), sometimes called gamma-Poisson distribution.

Some distributions have been specially named as compounds: beta-binomial distribution, Beta negative binomial distribution, gamma-normal distribution.

Examples:

  • If X is a Binomial(n,p) random variable, and parameter p is a random variable with beta(αあるふぁ, βべーた) distribution, then X is distributed as a Beta-Binomial(αあるふぁ,βべーた,n).
  • If X is a negative-binomial(r,p) random variable, and parameter p is a random variable with beta(αあるふぁ,βべーた) distribution, then X is distributed as a Beta negative binomial distribution(r,αあるふぁ,βべーた).

See also

edit

References

edit
  1. ^ LEEMIS, Lawrence M.; Jacquelyn T. MCQUESTON (February 2008). "Univariate Distribution Relationships" (PDF). American Statistician. 62 (1): 45–53. doi:10.1198/000313008x270448. S2CID 9367367.
  2. ^ Swat, MJ; Grenon, P; Wimalaratne, S (2016). "ProbOnto: ontology and knowledge base of probability distributions". Bioinformatics. 32 (17): 2719–21. doi:10.1093/bioinformatics/btw170. PMC 5013898. PMID 27153608.
  3. ^ Cook, John D. "Diagram of distribution relationships".
  4. ^ Dinov, Ivo D.; Siegrist, Kyle; Pearl, Dennis; Kalinin, Alex; Christou, Nicolas (2015). "Probability Distributome: a web computational infrastructure for exploring the properties, interrelations, and applications of probability distributions". Computational Statistics. 594 (2): 249–271. doi:10.1007/s00180-015-0594-6. PMC 4856044. PMID 27158191.
edit