Psychology Wiki
No edit summary
 
Line 1: Line 1:
 
{{StatPsy}}
 
{{StatPsy}}
   
  +
A '''probability distribution''' describes the ''values'' and ''[[probabilities]]'' that a [[random event]] can take place. The values must cover all of the possible outcomes of the event, while the total probabilities must [[sum]] to exactly 1, or 100%. For example, a single coin flip can take values ''Heads'' or ''Tails'' with a probability of exactly 1/2 for each; these two values and two probabilities make up the probability distribution of the single coin flipping event. This distribution is called a ''[[discrete probability distribution|discrete distribution]]'' because there are a [[Countable|countable]] number of discrete outcomes with positive probabilities.
In [[mathematics]] and [[statistics]], a '''probability distribution''', more properly called a '''probability distribution function''', assigns to every [[interval (mathematics)|interval]] of the [[real number]]s a [[probability]], so that the [[probability axioms]] are satisfied. In technical terms, a probability distribution is a [[probability measure]] whose domain is the [[Borel algebra]] on the reals.
 
   
  +
A ''[[continuous probability distribution|continuous distribution]]'' describes events over a continuous range, where the probability of a specific outcome is zero. For example, a dart thrown at a dartboard has essentially zero probability of landing at a specific point, since a point is [[infinitesimal|vanishingly small]], but it has some probability of landing within a given area. The probability of landing within the small area of the bullseye would (hopefully) be greater than landing on an equivalent area elsewhere on the board. A smooth function that describes the probability of landing anywhere on the dartboard is the probability distribution of the dart throwing event. The [[integral]] of the [[probability density function]] (pdf) over the entire area of the dartboard (and, perhaps, the wall surrounding it) must be equal to 1, since each dart must land somewhere.
A probability distribution is a special case of the more general notion of a [[probability measure]], which is a function that assigns probabilities satisfying the [[Kolmogorov axioms]] to the measurable sets of a [[measurable space]]. Additionally, some authors define a distribution generally as the probability measure induced by a [[random variable]] ''X'' on its [[range (mathematics)|range]] - the probability of a set ''B'' is <math>P(X^{-1}(B))</math>. However, this article discusses only probability measures over the real numbers.
 
   
  +
The concept of the probability distribution and the [[random variables]] which they describe underlies the mathematical discipline of [[probability theory]], and the science of [[statistics]]. There is spread or variability in almost any value that can be measured in a population (e.g. height of people, durability of a metal, etc.); almost all measurements are made with some [[measurement error|intrinsic error]]; in [[physics]] many processes are described probabilistically, from the [[kinetic theory|kinetic properties of gases]] to the [[quantum mechanical]] description of [[fundamental particles]]. For these and many other reasons, simple [[numbers]] are often inadequate for describing a quantity, while probability distributions are often more appropriate models. There are, however, considerable mathematical complications in manipulating probability distributions, since most standard [[arithmetic]] and [[algebraic]] manipulations cannot be applied.
==Formal definition==
 
Every random variable gives rise to a probability distribution, and this distribution contains most of the important information about the variable. If ''X'' is a random variable, the corresponding probability distribution assigns to the interval [''a'', ''b''] the probability Pr[''a'' ≤ ''X'' ≤ ''b''], i.e. the probability that the variable ''X'' will take a value in the interval [''a'', ''b''].
 
The probability distribution of the variable ''X'' can be uniquely described by its [[cumulative distribution function]] ''F''(''x''), which is defined by
 
   
  +
== Rigorous definitions ==
:<math> F(x) = \Pr\left[ X \le x \right] </math>
 
   
  +
In [[probability theory]], every [[random variable]] may be attributed to a function defined on a state space equipped with a '''probability distribution''' that assigns a [[probability]] to every [[subset]] (more precisely every measurable subset) of its [[probability space|state space]] in such a way that the [[probability axioms]] are satisfied. That is, probability distributions are [[probability measure]]s defined over a state space instead of the [[sample space]]. A random variable then defines a probability measure on the sample space by assigning a subset of the sample space the probability of its inverse image in the state space. In other words the probability distribution of a random variable is the [[push forward measure]] of the probability distribution on the state space.
for any ''x'' in '''R'''.
 
   
  +
===Probability distributions of real-valued random variables===
A distribution is called ''discrete'' if its cumulative distribution function consists of a sequence of finite jumps, which means that it belongs to a [[discrete random variable]] ''X'': a variable which can only attain values from a certain finite or [[countable]] set. By one convention, a distribution is called ''continuous'' if its cumulative distribution function is [[continuous function|continuous]], which means that it belongs to a random variable ''X'' for which Pr[ ''X'' = ''x'' ] = 0 for all ''x'' in '''R'''. Another convention reserves the term ''continuous probability distribution'' for [[absolute continuity|absolutely continuous]] distributions. These can be expressed by a [[probability density function]]: a non-negative [[Lebesgue integration|Lebesgue integrable]] function ''f'' defined on the real numbers such that
 
  +
Because a probability distribution Pr on the real line is determined by the probability of being in a half-open interval Pr<nowiki>(</nowiki>''a'',&nbsp;''b''<nowiki>]</nowiki>, the probability distribution of a real-valued random variable ''X'' is completely characterized by its [[cumulative distribution function]]:
   
 
:<math> F(x) = \Pr \left[ X \le x \right] \qquad \forall x \in \mathbb{R}.</math>
  +
  +
====Discrete probability distribution====
  +
{{main|Discrete probability distribution}}
 
A probability distribution is called ''discrete'' if its cumulative distribution function only increases in jumps.
  +
  +
The [[set]] of all values that a discrete random variable can assume with non-zero probability is either [[finite set|finite]] or [[countably infinite]] because the sum of uncountably many positive [[real number]]s (which is the smallest upper bound of the set of all finite partial sums) always diverges to infinity. Typically, the set of possible values is topologically discrete in the sense that all its points are [[isolated point]]s. But, there are discrete random variables for which this countable set is [[dense set|dense]] on the real line.
  +
  +
Discrete distributions are characterized by a [[probability mass function]], <math>p</math> such that
 
:<math>
 
:<math>
\Pr \left[ a \le X \le b \right] = \int_a^b f(x)\,dx
+
F(x) = \Pr \left[X \le x \right] = \sum_{x_i \le x} p(x_i).
 
</math>
 
</math>
   
  +
====Continuous probability distribution====
for all ''a'' and ''b''. Of course, discrete distributions do not admit such a density; there also exist some continuous distributions like the [[devil's staircase]] that do not admit a density.
 
  +
{{main|Continuous probability distribution}}
  +
By one convention, a probability distribution is called ''continuous'' if its cumulative distribution function is [[continuous function|continuous]], which means that it belongs to a random variable ''X'' for which Pr[ ''X'' = ''x'' ] = 0 for all ''x'' in '''R'''.
  +
  +
Another convention reserves the term ''continuous probability distribution'' for [[absolute continuity|absolutely continuous]] distributions. These distributions can be characterized by a [[probability density function]]: a non-negative [[Lebesgue integration|Lebesgue integrable]] function <math>f</math> defined on the real numbers such that
   
Discrete distribution function is expressed as -
 
 
:<math>
 
:<math>
F(x) = \Pr \left[X \le x \right] = \sum_{x_i \le x} p(x_i)
+
F(x) = \Pr \left[ X \le x \right] = \int_{-\infty}^x f(t)\,dt
 
</math>
 
</math>
for <math>i = 1, 2, ...\,\!</math>.
 
   
 
Discrete distributions and some continuous distributions (like the [[devil's staircase]]) do not admit such a density.
Here <math>p(x_i)\,\!</math> is called [[probability mass function]].
 
   
  +
===Terminology===
* The ''support'' of a distribution is the smallest closed set whose complement has probability zero.
+
The '''support''' of a distribution is the smallest closed set whose complement has probability zero.
* The probability distribution of the sum of two independent random variables is the [[convolution]] of each of their distributions.
 
* The probability distribution of the difference of two random variables is the [[cross-correlation]] of each of their distributions.
 
   
 
The probability distribution of the sum of two independent random variables is the '''[[convolution]]''' of each of their distributions.
== List of important probability distributions ==
 
   
 
The probability distribution of the difference of two random variables is the '''[[cross-correlation]]''' of each of their distributions.
Several probability distributions are so important in theory or applications that they have been given specific names:
 
  +
  +
A '''discrete random variable''' is a random variable whose probability distribution is discrete. Similarly, a '''continuous random variable''' is a random variable whose probability distribution is continuous.
  +
 
== List of important probability distributions ==
  +
{{splitsection}}
  +
Certain random variables occur very often in probability theory, in some cases due to their application to many natural and physical processes, and in some cases due to theoretical reasons such as the [[central limit theorem]], the [[Poisson limit theorem]], or properties such as [[memorylessness]] or other [[characterization (mathematics)|characterizations]]. Their distributions therefore have gained ''special importance'' in probability theory.
   
 
===Discrete distributions===
 
===Discrete distributions===
 
====With finite support====
 
====With finite support====
   
* The [[Bernoulli distribution]], which takes value 1 with probability ''p'' and value 0 with probability ''q'' = 1 &minus; ''p''.
+
*The [[Bernoulli distribution]], which takes value 1 with probability ''p'' and value 0 with probability ''q'' = 1 &minus; ''p''.
 
* The [[Rademacher distribution]], which takes value 1 with probability 1/2 and value &minus;1 with probability 1/2.
 
* The [[Rademacher distribution]], which takes value 1 with probability 1/2 and value &minus;1 with probability 1/2.
 
* The [[binomial distribution]] describes the number of successes in a series of independent Yes/No experiments.
 
* The [[binomial distribution]] describes the number of successes in a series of independent Yes/No experiments.
* The [[degenerate distribution]] at ''x''<sub>0</sub>, where ''X'' is certain to take the value ''x<sub>0</sub>''. This does not look random, but it satisfies the definition of [[random variable]]. This is useful because it puts deterministic variables and random variables in the same formalism.
+
* The [[degenerate distribution]] at ''x''<sub>0</sub>, where ''X'' is certain to take the value ''x<sub>0</sub>''. This does not look random, but it satisfies the definition of [[random variable]]. It is useful because it puts deterministic variables and random variables in the same formalism.
 
* The [[Uniform distribution (discrete)|discrete uniform distribution]], where all elements of a finite [[set theory|set]] are equally likely. This is supposed to be the distribution of a balanced coin, an unbiased die, a casino roulette or a well-shuffled deck. Also, one can use measurements of quantum states to generate uniform random variables. All these are "physical" or "mechanical" devices, subject to design flaws or perturbations, so the uniform distribution is only an approximation of their behaviour. In digital computers, [[Pseudorandom number sequence|pseudo-random number generators]] are used to produce a [[randomness|statistically random]] discrete uniform distribution.
 
* The [[Uniform distribution (discrete)|discrete uniform distribution]], where all elements of a finite [[set theory|set]] are equally likely. This is supposed to be the distribution of a balanced coin, an unbiased die, a casino roulette or a well-shuffled deck. Also, one can use measurements of quantum states to generate uniform random variables. All these are "physical" or "mechanical" devices, subject to design flaws or perturbations, so the uniform distribution is only an approximation of their behaviour. In digital computers, [[Pseudorandom number sequence|pseudo-random number generators]] are used to produce a [[randomness|statistically random]] discrete uniform distribution.
 
* The [[hypergeometric distribution]], which describes the number of successes in the first ''m'' of a series of ''n'' Yes/No experiments, if the total number of successes is known.
 
* The [[hypergeometric distribution]], which describes the number of successes in the first ''m'' of a series of ''n'' Yes/No experiments, if the total number of successes is known.
Line 59: Line 74:
 
[[Image:Poisson distribution PMF.png|150px|thumb|[[Poisson distribution]]]]
 
[[Image:Poisson distribution PMF.png|150px|thumb|[[Poisson distribution]]]]
 
* The [[logarithmic distribution|logarithmic (series) distribution]]
 
* The [[logarithmic distribution|logarithmic (series) distribution]]
* The [[negative binomial distribution]], a generalization of the geometric distribution to the ''n''th success.
+
* The [[negative binomial distribution]], a generalization of the geometric distribution to the ''n''th success
 
* The [[parabolic fractal distribution]]
 
* The [[parabolic fractal distribution]]
 
* The [[Poisson distribution]], which describes a very large number of individually unlikely events that happen in a certain time interval.
 
* The [[Poisson distribution]], which describes a very large number of individually unlikely events that happen in a certain time interval.
  +
* The [[Conway-Maxwell-Poisson distribution]], a generalization of the Poisson distribution with an adjustable rate of decay
 
[[Image:SkellamDistribution.png|150px|thumb|[[Skellam distribution]]]]
 
[[Image:SkellamDistribution.png|150px|thumb|[[Skellam distribution]]]]
* The [[Skellam distribution]], the distribution of the difference between two independent Poisson-distributed random variables.
+
* The [[Skellam distribution]], the distribution of the difference between two independent Poisson-distributed random variables
 
* The [[Yule-Simon distribution]]
 
* The [[Yule-Simon distribution]]
 
* The [[zeta distribution]] has uses in applied statistics and statistical mechanics, and perhaps may be of interest to number theorists. It is the [[Zipf distribution]] for an infinite number of elements.
 
* The [[zeta distribution]] has uses in applied statistics and statistical mechanics, and perhaps may be of interest to number theorists. It is the [[Zipf distribution]] for an infinite number of elements.
Line 75: Line 91:
 
** The [[rectangular distribution]] is a uniform distribution on [-1/2,1/2].
 
** The [[rectangular distribution]] is a uniform distribution on [-1/2,1/2].
 
* The [[Dirac delta function]] although not strictly a function, is a limiting form of many continuous probability functions. It represents a ''discrete'' probability distribution concentrated at 0 &mdash; a [[degenerate distribution]] &mdash; but the notation treats it as if it were a continuous distribution.
 
* The [[Dirac delta function]] although not strictly a function, is a limiting form of many continuous probability functions. It represents a ''discrete'' probability distribution concentrated at 0 &mdash; a [[degenerate distribution]] &mdash; but the notation treats it as if it were a continuous distribution.
  +
* The [[Kent distribution]] on the three-dimensional sphere
 
* The [[Kumaraswamy distribution]] is as versatile as the Beta distribution but has simple closed forms for both the cdf and the pdf.
 
* The [[Kumaraswamy distribution]] is as versatile as the Beta distribution but has simple closed forms for both the cdf and the pdf.
 
* The [[logarithmic distribution (continuous)]]
 
* The [[logarithmic distribution (continuous)]]
 
* The [[triangular distribution]] on [''a'', ''b''], a special case of which is the distribution of the sum of two uniformly distributed random variables (the ''convolution'' of two uniform distributions).
 
* The [[triangular distribution]] on [''a'', ''b''], a special case of which is the distribution of the sum of two uniformly distributed random variables (the ''convolution'' of two uniform distributions).
* The [[von Mises distribution]] on the circle.
+
* The [[truncated normal distribution]] on [''a'', ''b'']
* The [[von Mises-Fisher distribution]] on the N-dimensional sphere has the [[von Mises distribution]]
+
* The [[U-quadratic distribution]] on [''a'', ''b'']
 
* The [[von Mises distribution]] on the circle
as a special case.
 
* The [[Kent distribution]] on the three-dimensional sphere.
+
* The [[von Mises-Fisher distribution]] on the N-dimensional sphere has the [[von Mises distribution]] as a special case.
 
*The [[Wigner semicircle distribution]] is important in the theory of [[random matrices]].
 
*The [[Wigner semicircle distribution]] is important in the theory of [[random matrices]].
   
Line 100: Line 117:
 
** The [[Erlang distribution]], which is a special case of the gamma distribution with integral shape parameter, developed to predict waiting times in [[queuing systems]].
 
** The [[Erlang distribution]], which is a special case of the gamma distribution with integral shape parameter, developed to predict waiting times in [[queuing systems]].
 
** The [[inverse-gamma distribution]]
 
** The [[inverse-gamma distribution]]
  +
* The [[folded normal distribution]]
 
* The [[half-normal distribution]]
 
* The [[half-normal distribution]]
  +
* The [[inverse Gaussian distribution]], also known as the Wald distribution
 
* The [[Lévy distribution]]
 
* The [[Lévy distribution]]
 
* The [[log-logistic distribution]]
 
* The [[log-logistic distribution]]
Line 106: Line 125:
 
[[Image:Pareto distributionPDF.png|thumb|150px|[[Pareto distribution]]]]
 
[[Image:Pareto distributionPDF.png|thumb|150px|[[Pareto distribution]]]]
 
* The [[Pareto distribution]], or "power law" distribution, used in the analysis of financial data and critical behavior.
 
* The [[Pareto distribution]], or "power law" distribution, used in the analysis of financial data and critical behavior.
* The [[Pearson Type III distribution]] (see [[Pearson distribution]]s)
+
* The Pearson Type III distribution (see [[Pearson distribution]]s)
 
* The [[Rayleigh distribution]]
 
* The [[Rayleigh distribution]]
 
* The [[Rayleigh mixture distribution]]
 
* The [[Rayleigh mixture distribution]]
 
* The [[Rice distribution]]
 
* The [[Rice distribution]]
 
* The [[type-2 Gumbel distribution]]
 
* The [[type-2 Gumbel distribution]]
 
* The [[Weibull distribution]] or Rosin Rammler distribution, of which the exponential distribution is a special case, is used to model the lifetime of technical devices and used to describe the [[particle size distribution]] of particles generated by [[grinding]], [[milling]] and [[crushing]] operations.
* The [[Wald distribution]]
 
* The [[Weibull distribution]], of which the exponential distribution is a special case, is used to model the lifetime of technical devices.
 
   
 
====Supported on the whole real line====
 
====Supported on the whole real line====
Line 139: Line 157:
 
===Joint distributions===
 
===Joint distributions===
   
For any set of [[Statistical independence|independent]] random variables the [[probability density function]] of the joint distribution is the product of the individual ones.
+
For any set of [[Statistical independence|independent]] random variables the [[probability density function]] of their [[joint distribution]] is the product of their individual density functions.
   
 
====Two or more random variables on the same sample space====
 
====Two or more random variables on the same sample space====
Line 145: Line 163:
 
* [[Dirichlet distribution]], a generalization of the [[beta distribution]].
 
* [[Dirichlet distribution]], a generalization of the [[beta distribution]].
 
*The [[Ewens's sampling formula]] is a probability distribution on the set of all [[integer partition|partitions of an integer]] ''n'', arising in [[population genetics]].
 
*The [[Ewens's sampling formula]] is a probability distribution on the set of all [[integer partition|partitions of an integer]] ''n'', arising in [[population genetics]].
* [[Balding-Nichols Model]]
+
* [[Balding-Nichols model]]
 
* [[multinomial distribution]], a generalization of the [[binomial distribution]].
 
* [[multinomial distribution]], a generalization of the [[binomial distribution]].
 
* [[multivariate normal distribution]], a generalization of the [[normal distribution]].
 
* [[multivariate normal distribution]], a generalization of the [[normal distribution]].
Line 159: Line 177:
   
 
* The [[Cantor distribution]]
 
* The [[Cantor distribution]]
  +
* [[Phase-type distribution]]
 
* [[Truncated distribution]]
 
* [[Truncated distribution]]
   
== See also ==
 
   
  +
 
== See also ==
 
*[[random variable]]
 
*[[copula (statistics)]]
 
*[[copula (statistics)]]
 
*[[cumulative distribution function]]
 
*[[cumulative distribution function]]
Line 168: Line 189:
 
*[[list of statistical topics]]
 
*[[list of statistical topics]]
 
*[[probability density function]]
 
*[[probability density function]]
*[[random variable]]
 
 
*[[histogram]]
 
*[[histogram]]
  +
*[[Inverse transform sampling]]
  +
*[[Riemann-Stieltjes integral#Application to probability theory|Riemann-Stieltjes integral: Application to probability theory]]
   
 
{{ProbDistributions}}
 
{{ProbDistributions}}
  +
{{Statistics}}
   
 
==External links==
 
==External links==
  +
{{commons|Probability distribution|Probability distribution}}
 
  +
*[http://stattrek.com/Lesson2/ProbabilityDistribution.aspx?Tutorial=Stat Probability Distributions for Beginners.]
 
*[http://www.socr.ucla.edu/htmls/SOCR_Distributions.html Interactive Discrete and Continuous Probability Distributions]
 
*[http://www.socr.ucla.edu/htmls/SOCR_Distributions.html Interactive Discrete and Continuous Probability Distributions]
 
*[http://www.causascientia.org/math_stat/Dists/Compendium.pdf A Compendium of Common Probability Distributions]
 
*[http://www.causascientia.org/math_stat/Dists/Compendium.pdf A Compendium of Common Probability Distributions]
 
*[http://www.xycoon.com/contdistroverview.htm Statistical Distributions - Overview]
 
*[http://www.xycoon.com/contdistroverview.htm Statistical Distributions - Overview]
*[http://www.stat.vt.edu/~sundar/java/applets/Distributions.html Probability Distributions]
 
*[http://www.brighton-webs.co.uk/distributions/ Statistics - Distributions]
 
 
*[http://www.sitmo.com/eqcat/8 Probability Distributions] in Quant Equation Archive, sitmo
 
*[http://www.sitmo.com/eqcat/8 Probability Distributions] in Quant Equation Archive, sitmo
  +
*[http://www.covariable.com/continuous.html A Probability Distribution Calculator]
  +
 
[[Category:Probability and statistics]]
 
[[Category:Probability and statistics]]
[[Category:Probability distributions| ]]
+
[[Category:Probability distributions|*]]
 
   
  +
<!--
  +
[[ar:توزيع احتمالي]]
  +
[[cs:Rozdělení pravděpodobnosti]]
  +
[[de:Wahrscheinlichkeitsverteilung]]
  +
[[es:Distribución de probabilidad]]
  +
[[eo:Probabla distribuo]]
  +
[[fa:توزیع احتمال]]
  +
[[fr:Loi de probabilité]]
  +
[[gl:Distribución de probabilidade]]
  +
[[it:Distribuzione di probabilità]]
  +
[[he:התפלגות]]
  +
[[lt:Skirstinys]]
  +
[[nl:Kansverdeling]]
  +
[[ja:確率分布]]
  +
[[pl:Rozkład zmiennej losowej]]
  +
[[pt:Distribuição de probabilidade]]
  +
[[ru:Распределение вероятностей]]
  +
[[su:Sebaran probabilitas]]
  +
[[sv:Sannolikhetsfördelning]]
  +
[[vi:Phân bố xác suất]]
  +
[[zh:概率分布]]
  +
-->
   
 
{{enWP|Probability Distribution}}
 
{{enWP|Probability Distribution}}

Latest revision as of 23:02, 16 April 2008

Assessment | Biopsychology | Comparative | Cognitive | Developmental | Language | Individual differences | Personality | Philosophy | Social |
Methods | Statistics | Clinical | Educational | Industrial | Professional items | World psychology |

Statistics: Scientific method · Research methods · Experimental design · Undergraduate statistics courses · Statistical tests · Game theory · Decision theory


A probability distribution describes the values and probabilities that a random event can take place. The values must cover all of the possible outcomes of the event, while the total probabilities must sum to exactly 1, or 100%. For example, a single coin flip can take values Heads or Tails with a probability of exactly 1/2 for each; these two values and two probabilities make up the probability distribution of the single coin flipping event. This distribution is called a discrete distribution because there are a countable number of discrete outcomes with positive probabilities.

A continuous distribution describes events over a continuous range, where the probability of a specific outcome is zero. For example, a dart thrown at a dartboard has essentially zero probability of landing at a specific point, since a point is vanishingly small, but it has some probability of landing within a given area. The probability of landing within the small area of the bullseye would (hopefully) be greater than landing on an equivalent area elsewhere on the board. A smooth function that describes the probability of landing anywhere on the dartboard is the probability distribution of the dart throwing event. The integral of the probability density function (pdf) over the entire area of the dartboard (and, perhaps, the wall surrounding it) must be equal to 1, since each dart must land somewhere.

The concept of the probability distribution and the random variables which they describe underlies the mathematical discipline of probability theory, and the science of statistics. There is spread or variability in almost any value that can be measured in a population (e.g. height of people, durability of a metal, etc.); almost all measurements are made with some intrinsic error; in physics many processes are described probabilistically, from the kinetic properties of gases to the quantum mechanical description of fundamental particles. For these and many other reasons, simple numbers are often inadequate for describing a quantity, while probability distributions are often more appropriate models. There are, however, considerable mathematical complications in manipulating probability distributions, since most standard arithmetic and algebraic manipulations cannot be applied.

Rigorous definitions

In probability theory, every random variable may be attributed to a function defined on a state space equipped with a probability distribution that assigns a probability to every subset (more precisely every measurable subset) of its state space in such a way that the probability axioms are satisfied. That is, probability distributions are probability measures defined over a state space instead of the sample space. A random variable then defines a probability measure on the sample space by assigning a subset of the sample space the probability of its inverse image in the state space. In other words the probability distribution of a random variable is the push forward measure of the probability distribution on the state space.

Probability distributions of real-valued random variables

Because a probability distribution Pr on the real line is determined by the probability of being in a half-open interval Pr(ab], the probability distribution of a real-valued random variable X is completely characterized by its cumulative distribution function:

Discrete probability distribution

Main article: Discrete probability distribution

A probability distribution is called discrete if its cumulative distribution function only increases in jumps.

The set of all values that a discrete random variable can assume with non-zero probability is either finite or countably infinite because the sum of uncountably many positive real numbers (which is the smallest upper bound of the set of all finite partial sums) always diverges to infinity. Typically, the set of possible values is topologically discrete in the sense that all its points are isolated points. But, there are discrete random variables for which this countable set is dense on the real line.

Discrete distributions are characterized by a probability mass function, such that

Continuous probability distribution

Main article: Continuous probability distribution

By one convention, a probability distribution is called continuous if its cumulative distribution function is continuous, which means that it belongs to a random variable X for which Pr[ X = x ] = 0 for all x in R.

Another convention reserves the term continuous probability distribution for absolutely continuous distributions. These distributions can be characterized by a probability density function: a non-negative Lebesgue integrable function defined on the real numbers such that

Discrete distributions and some continuous distributions (like the devil's staircase) do not admit such a density.

Terminology

The support of a distribution is the smallest closed set whose complement has probability zero.

The probability distribution of the sum of two independent random variables is the convolution of each of their distributions.

The probability distribution of the difference of two random variables is the cross-correlation of each of their distributions.

A discrete random variable is a random variable whose probability distribution is discrete. Similarly, a continuous random variable is a random variable whose probability distribution is continuous.

List of important probability distributions

Certain random variables occur very often in probability theory, in some cases due to their application to many natural and physical processes, and in some cases due to theoretical reasons such as the central limit theorem, the Poisson limit theorem, or properties such as memorylessness or other characterizations. Their distributions therefore have gained special importance in probability theory.

Discrete distributions

With finite support

  • The Bernoulli distribution, which takes value 1 with probability p and value 0 with probability q = 1 − p.
  • The Rademacher distribution, which takes value 1 with probability 1/2 and value −1 with probability 1/2.
  • The binomial distribution describes the number of successes in a series of independent Yes/No experiments.
  • The degenerate distribution at x0, where X is certain to take the value x0. This does not look random, but it satisfies the definition of random variable. It is useful because it puts deterministic variables and random variables in the same formalism.
  • The discrete uniform distribution, where all elements of a finite set are equally likely. This is supposed to be the distribution of a balanced coin, an unbiased die, a casino roulette or a well-shuffled deck. Also, one can use measurements of quantum states to generate uniform random variables. All these are "physical" or "mechanical" devices, subject to design flaws or perturbations, so the uniform distribution is only an approximation of their behaviour. In digital computers, pseudo-random number generators are used to produce a statistically random discrete uniform distribution.
  • The hypergeometric distribution, which describes the number of successes in the first m of a series of n Yes/No experiments, if the total number of successes is known.
  • Zipf's law or the Zipf distribution. A discrete power-law distribution, the most famous example of which is the description of the frequency of words in the English language.
  • The Zipf-Mandelbrot law is a discrete power law distribution which is a generalization of the Zipf distribution.

With infinite support

  • The Boltzmann distribution, a discrete distribution important in statistical physics which describes the probabilities of the various discrete energy levels of a system in thermal equilibrium. It has a continuous analogue. Special cases include:
  • The geometric distribution, a discrete distribution which describes the number of attempts needed to get the first success in a series of independent Yes/No experiments.
Poisson distribution PMF

Poisson distribution

  • The logarithmic (series) distribution
  • The negative binomial distribution, a generalization of the geometric distribution to the nth success
  • The parabolic fractal distribution
  • The Poisson distribution, which describes a very large number of individually unlikely events that happen in a certain time interval.
  • The Conway-Maxwell-Poisson distribution, a generalization of the Poisson distribution with an adjustable rate of decay
File:SkellamDistribution.png

Skellam distribution

  • The Skellam distribution, the distribution of the difference between two independent Poisson-distributed random variables
  • The Yule-Simon distribution
  • The zeta distribution has uses in applied statistics and statistical mechanics, and perhaps may be of interest to number theorists. It is the Zipf distribution for an infinite number of elements.

Continuous distributions

Supported on a bounded interval

File:Beta distribution pdf.png

Beta distribution

  • The Beta distribution on [0,1], of which the uniform distribution is a special case, and which is useful in estimating success probabilities.
File:Uniform distribution PDF.png

continuous uniform distribution

  • The continuous uniform distribution on [a,b], where all points in a finite interval are equally likely.
    • The rectangular distribution is a uniform distribution on [-1/2,1/2].
  • The Dirac delta function although not strictly a function, is a limiting form of many continuous probability functions. It represents a discrete probability distribution concentrated at 0 — a degenerate distribution — but the notation treats it as if it were a continuous distribution.
  • The Kent distribution on the three-dimensional sphere
  • The Kumaraswamy distribution is as versatile as the Beta distribution but has simple closed forms for both the cdf and the pdf.
  • The logarithmic distribution (continuous)
  • The triangular distribution on [a, b], a special case of which is the distribution of the sum of two uniformly distributed random variables (the convolution of two uniform distributions).
  • The truncated normal distribution on [a, b]
  • The U-quadratic distribution on [a, b]
  • The von Mises distribution on the circle
  • The von Mises-Fisher distribution on the N-dimensional sphere has the von Mises distribution as a special case.
  • The Wigner semicircle distribution is important in the theory of random matrices.

Supported on semi-infinite intervals, usually [0,∞)

Chi-square distributionPDF

chi-square distribution

Exponential distribution pdf

Exponential distribution

  • The exponential distribution, which describes the time between consecutive rare random events in a process with no memory.
  • The F-distribution, which is the distribution of the ratio of two (normalized) chi-square distributed random variables, used in the analysis of variance. (Called the beta prime distribution when it is the ratio of two chi-square variates which are not normalized by dividing them by their numbers of degrees of freedom.)
Gamma distribution pdf

Gamma distribution

  • The Gamma distribution, which describes the time until n consecutive rare random events occur in a process with no memory.
    • The Erlang distribution, which is a special case of the gamma distribution with integral shape parameter, developed to predict waiting times in queuing systems.
    • The inverse-gamma distribution
  • The folded normal distribution
  • The half-normal distribution
  • The inverse Gaussian distribution, also known as the Wald distribution
  • The Lévy distribution
  • The log-logistic distribution
  • The log-normal distribution, describing variables which can be modelled as the product of many small independent positive variables.
Pareto distributionPDF

Pareto distribution

  • The Pareto distribution, or "power law" distribution, used in the analysis of financial data and critical behavior.
  • The Pearson Type III distribution (see Pearson distributions)
  • The Rayleigh distribution
  • The Rayleigh mixture distribution
  • The Rice distribution
  • The type-2 Gumbel distribution
  • The Weibull distribution or Rosin Rammler distribution, of which the exponential distribution is a special case, is used to model the lifetime of technical devices and used to describe the particle size distribution of particles generated by grinding, milling and crushing operations.

Supported on the whole real line

File:Cauchy distribution pdf.png

Cauchy distribution

File:Laplace distribution pdf.png

Laplace distribution

File:LevyDistribution.png

Levy distribution

Normal distribution pdf

Normal distribution

  • The Cauchy distribution, an example of a distribution which does not have an expected value or a variance. In physics it is usually called a Lorentzian profile, and is associated with many processes, including resonance energy distribution, impact and natural spectral line broadening and quadratic stark line broadening.
  • The Fisher-Tippett, extreme value, or log-Weibull distribution
    • The Gumbel distribution, a special case of the Fisher-Tippett distribution
  • Fisher's z-distribution
  • The generalized extreme value distribution
  • The hyperbolic distribution
  • The hyperbolic secant distribution
  • The Landau distribution
  • The Laplace distribution
  • The Lévy skew alpha-stable distribution is often used to characterize financial data and critical behavior.
  • The map-Airy distribution
  • The normal distribution, also called the Gaussian or the bell curve. It is ubiquitous in nature and statistics due to the central limit theorem: every variable that can be modelled as a sum of many small independent variables is approximately normal.
  • The Pearson Type IV distribution (see Pearson distributions)
  • Student's t-distribution, useful for estimating unknown means of Gaussian populations.
    • The noncentral t-distribution
  • The type-1 Gumbel distribution
  • The Voigt distribution, or Voigt profile, is the convolution of a normal distribution and a Cauchy distribution. It is found in spectroscopy when spectral line profiles are broadened by a mixture of Lorentzian and Doppler broadening mechanisms.

Joint distributions

For any set of independent random variables the probability density function of their joint distribution is the product of their individual density functions.

Two or more random variables on the same sample space

Matrix-valued distributions

  • Wishart distribution
  • matrix normal distribution
  • matrix t-distribution
  • Hotelling's T-square distribution

Miscellaneous distributions

  • The Cantor distribution
  • Phase-type distribution
  • Truncated distribution


See also

Bvn-small Probability distributions [[[:Template:Tnavbar-plain-nodiv]]]
Univariate Multivariate
Discrete: BernoullibinomialBoltzmanncompound PoissondegeneratedegreeGauss-Kuzmingeometrichypergeometriclogarithmicnegative binomialparabolic fractalPoissonRademacherSkellamuniformYule-SimonzetaZipfZipf-Mandelbrot Ewensmultinomial
Continuous: BetaBeta primeCauchychi-squareDirac delta functionErlangexponentialexponential powerFfadingFisher's zFisher-TippettGammageneralized extreme valuegeneralized hyperbolicgeneralized inverse GaussianHotelling's T-squarehyperbolic secanthyper-exponentialhypoexponentialinverse chi-squareinverse gaussianinverse gammaKumaraswamyLandauLaplaceLévyLévy skew alpha-stablelogisticlog-normalMaxwell-BoltzmannMaxwell speednormal (Gaussian)ParetoPearsonpolarraised cosineRayleighrelativistic Breit-WignerRiceStudent's ttriangulartype-1 Gumbeltype-2 GumbeluniformVoigtvon MisesWeibullWigner semicircle DirichletKentmatrix normalmultivariate normalvon Mises-FisherWigner quasiWishart
Miscellaneous: Cantorconditionalexponential familyinfinitely divisiblelocation-scale familymarginalmaximum entropy phase-typeposterior priorquasisampling




External links


This page uses Creative Commons Licensed content from Wikipedia (view authors).