Assessment |
Biopsychology |
Comparative |
Cognitive |
Developmental |
Language |
Individual differences |
Personality |
Philosophy |
Social |
Methods |
Statistics |
Clinical |
Educational |
Industrial |
Professional items |
World psychology |
Statistics: Scientific method · Research methods · Experimental design · Undergraduate statistics courses · Statistical tests · Game theory · Decision theory
Probability density function PDF of the uniform probability distribution using the maximum convention at the transition points. Using maximum convention | |
Cumulative distribution function CDF of the uniform probability distribution. | |
Parameters | |
Support | |
cdf | |
Mean | |
Median | |
Mode | any value in |
Variance | |
Skewness | |
Kurtosis | |
Entropy | |
mgf | |
Char. func. |
In mathematics, the continuous uniform distributions are probability distributions such that all intervals of the same length are equally probable.
The continuous uniform distribution is a generalization of the rectangle function because of the shape of its probability density function. It is parametrised by the smallest and largest values that the uniformly-distributed random variable can take, a and b. The probability density function of the uniform distribution is thus:
The values at the two boundaries a and b are usually unimportant because they do not alter the values of the integrals of f(x) dx over any interval, nor of x f(x) dx or the like. Sometimes they are chosen to be zero, and sometimes chosen to be 1/(b − a). The latter is appropriate in the context of estimation by the method of maximum likelihood. In the context of Fourier analysis, one may take the value of f(a) or f(b) to be 1/(2(b − a)), since then the inverse transform of many integral transforms of this uniform function will yield back the function itself, rather than a function which is equal "almost everywhere", i.e. except at a finite number of points. Also, it is consistent with the sign function which has no such ambiguity.
The uniform distribution is normalized:
The cumulative distribution function Edit
The cumulative distribution function is:
The moment-generating function Edit
The moment-generating function is
from which we may calculate the raw moments m _{k}
For a random variable following this distribution, the expected value is then m_{1} = (a + b)/2 and the variance is m_{2} − m_{1}^{2} = (b − a)^{2}/12.
This distribution can be generalized to more complicated sets than intervals. If S is a Borel set of positive, finite measure, the uniform probability distribution on S can be specified by saying that the pdf is zero outside S and constantly equal to 1/K on S, where K is the Lebesgue measure of S.
For n ≥ 2, the nth cumulant of the uniform distribution on the interval [0, 1] is b_{b</sup>/n, where bn is the nth Bernoulli number. }
Standard uniform Edit
Restricting and , the resulting distribution is called a standard uniform distribution.
One interesting property of the standard uniform distribution is that if U_{1} is uniformly distributed,
then so is 1-U_{1}:
Related distributionsEdit
- is an exponential distribution if and .
Relationship to other functions Edit
As long as the same conventions are followed at the transition points, the probability density function may also be expressed in terms of the Heaviside step function:
and in terms of the rectangle function
There is no ambiguity at the transition point of the sign function. Using the half-maximum convention at the transition points, the uniform distribution may be expressed in terms of the sign function as:
The standard uniform distribution is the continuous uniform distribution with the values of a and b set to 0 and 1 respectively, so that the random variable can take values only between 0 and 1.
Sampling from a uniform distribution Edit
When working with probability, it is often useful to run experiments such as computational simulations. Many programming languages have the ability to generate pseudo-random numbers which are effectively distributed according to the standard uniform distribution.
If u is a value sampled from the standard uniform distribution, then the value a + (b − a)u follows the uniform distribution parametrised by a and b, as described above. Other transformations can be used to generate other statistical distributions from the uniform distribution. (see uses below)
Order statistcs Edit
Let X_{1}, ..., X_{n} be an i.i.d. sample from the uniform distribution on the interval [0, 1]. Let X_{(k)} (with the subscript k enclosed in parentheses) be the kth order statistic from this sample. Then the probability distribution of X_{(k)} is a Beta distribution with parameters k and n − k + 1. The expected value is
This fact is relevant to Q-Q plots.
Uses of the uniform distribution Edit
In statistics, when a p-value is used as a test statistic for a simple null hypothesis, and the distribution of the test statistic is continuous, then the test statistic is uniformly distributed between 0 and 1 if the null hypothesis is true.
Although the uniform distribution is not commonly found in nature, it is particularly useful for sampling from arbitrary distributions.
A general method is the inverse transform sampling method, which uses the cumulative distribution function (CDF) of the target random variable. This method is very useful in theoretical work. Since simulations using this method require inverting the CDF of the target variable, alternative methods have been devised for the cases where the CDF is not known in closed form. One such method is rejection sampling.
The normal distribution is an important example where the inverse transform method is not efficient. However, there is an exact method, the Box-Muller transformation, which uses the inverse transform to convert two independent uniform random variables into two independent normally distributed random variables.de:Stetige Gleichverteilungnl:Uniforme verdelingfi:Tasajakauma
This page uses Creative Commons Licensed content from Wikipedia (view authors). |