Normal Distribution
Definition
The probability density function for a normal (Gaussian) random variable \(X\) is given by:
where \(\mu\) is the mean and \(\sigma^2\) is the variance of \(X\). Gaussian distribution is denoted as \(X \sim \mathcal{N}(\mu, \sigma^2)\). The CDF is:
Normal distribution can be standardized to \(Z\), where \(Z \sim \mathcal{N}(0, 1)\):
Suppose we are interested in the probability that a random variable \(X\) distributed as \(\mathcal{N}(\mu, \sigma^2)\) falls between two bounds \(a\) and \(b\), \(\mathbb{P}(a < X < b)\). It is irrelevant whether the bounds are included or not since the normal distribution is continuous and \(\mathbb{P}(a < X < b) = \mathbb{P}(a \leq X \leq b)\). We have:
Note that when the bounds are infinite, since \(\Phi\) is a CDF:
Sigma Rules
Sigma rules (also known as empirical rule) state that for any normal distribution, the probability that an observation will fall in the interval \(\mu \pm k\sigma\) for \(k = 1, 2, 3\) is \(68.27\%\), \(95.45\%\), and \(99.73\%\), respectively. More precisely:
Mean and Variance
Recall that the moment-generating function is:
Then the moment-generating function for Gaussian distribution is:
Standarize the distribution via defining \(z = \frac{x - \mu}{\sigma}\), which implies \(x = z \sigma + \mu\). Then, using the change-of-variables:
Then the expected value of the Bernoulli distribution is:
The variance is the central second moment:
Combining Normal Random Variables
Any linear combination of independent normal variables is also normally distributed. Thus, only the mean and the variance of the variables involved in the linear combination need to be tracked. Let \(X_1, X_2, \ldots, X_n\) be independent normal random variables such that \(X_i \sim \mathcal{N}(\mu_i, \sigma^2_i)\). Then for \(a_i \in \mathbb{R}\):
where: