Probability density function
The red curve is the standard normal distribution |
|
Cumulative distribution function
|
|
Notation | |
---|---|
Parameters |
μ ∈ R — mean (location) σ2 > 0 — variance (squared scale) |
Support | x ∈ R |
CDF | |
Quantile | |
Mean | μ |
Median | μ |
Mode | μ |
Variance | |
Skewness | 0 |
Ex. kurtosis | 0 |
Entropy | |
MGF | |
CF | |
Fisher information |
In probability theory, the normal (or Gaussian) distribution is a very common continuous probability distribution. Normal distributions are important in statistics and are often used in the natural and social sciences to represent real-valued random variables whose distributions are not known.
The normal distribution is useful because of the central limit theorem. In its most general form, under some conditions (which include finite variance), it states that averages of random variables independently drawn from independent distributions converge in distribution to the normal, that is, become normally distributed when the number of random variables is sufficiently large. Physical quantities that are expected to be the sum of many independent processes (such as measurement errors) often have distributions that are nearly normal. Moreover, many results and methods (such as propagation of uncertainty and least squares parameter fitting) can be derived analytically in explicit form when the relevant variables are normally distributed.
The normal distribution is sometimes informally called the bell curve. However, many other distributions are bell-shaped (such as the Cauchy, Student's t, and logistic distributions). The terms Gaussian function and Gaussian bell curve are also ambiguous because they sometimes refer to multiples of the normal distribution that cannot be directly interpreted in terms of probabilities.