Normal Distribution

The Normal distribution (or gaussian distribution) is defined between and . It is parametrized by mean μ and variance σ, XN(μ,σ2) fX(x)=12πσ2e(xμ)22σ2 As already described, E[X]=μVar(X)=σ2

Standard Normal Distribution

A Standard Normal is defined as a normal distribution with μ=0 and σ2=1 Any normal distribution can be converted to a standard normal as X=Xμσ If Y=aX+b, then YN(aμ+b,a2σ2).

For a standard normal variable Z, it is standard to denote P(Zz)=Φ(z)P(Z=z)=ϕ(z)=N(0,1)

Further, for a given α(0,1), define zα by P(Z>zα)=α=1Φ(zα) Some standard values of α can be useful

  • z0.01=1.2816

  • z0.05=1.645

  • z0.025=1.96

  • z0.01=2.33

The 68–95–99.7 rule is also useful which states that

  • Probability of X lying between 1 standard deviation on either side of mean is 68%
  • Probability of X lying between 2 standard deviation on either side of mean is 95%
  • Probability of X lying between 3 standard deviation on either side of mean is 99.7%

It is often easier to derive all formulae here by first considering the standard normal and defining it completely. Then any normal distribution with mean μ and variance σ2 can be obtained using the transformation Y=σZ+μ. All calculations for mean, variance and mgf readily follow.

Moment Generating Function

E[etX]=etx12πσ2e(xμ)22σ2 We will rearrange the terms to form a perfect square in the exponent part with a changed mean. E[etX]=etx12πσ2e(xμ)22σ2=12πσ2exp((x(μ+σ2t))2(μ+σ2t)2+μ22σ2)=12πσ2exp((x(μ+σ2t)22σ2)exp(μt+σ2t22)E[etX]=exp(μt+σ2t22) Since the total integral of a normal distribution is 1 (the total probability).

Sum of Normal Distributions

With the help of moment generating functions, this calculation becomes easier. Let X1,Xn be n independent normal distributions with XiN(μi,σi2). E[et(X1+X2++Xn)]=i=1nE[etXi]=i=1neμit+σi2t22=exp(i=1nμit+i=1nσ2t22)X1+X2++XnN(μ1++μ2,σ12++σ22)

Multivariate Normal Distribution

Multivariate normal distribution is an extension of a normal distribution into multiple dimensions fX(x)=1(2π)d|Σ|exp(12(xμ)Σ1(xμ)T) for d dimensional vector x with mean vector μ and covriance matrix Σ.

This is easy to derive if we start from a standard normal first.

Consider n independent identically distributed standard normals ZiN(0,1). The joint distribution is fZ(z)=i=1nfZi(zi)=(12π)nexp(12i=1nzi2)=(12π)n2exp(12zTz)

E[Z]=0=E[Z]T (element wise). Cov(Z,Z)=E[ZZT]E[Z]E[Z]T=InasE[ZZT]ij=Cov(Zi,Zj)=0independenceE[ZZT]ii=Cov(Zi,Zi)=Var(Zi)=1independence

The mgf M(t)=E[etTZ]=exp(12tTt) by independence.

Now, assume a real symmetric semippositive definite matrix Σ that represents a variance-covariance matrix. From spectral decomposition, there exists Σ1/2 such that Σ1/2Σ1/2=Σ (from linear algebra). Also, this matrix is itself symmetric.

Assume a constant vector μ. Then, X=Σ1/2Z+μE[X]=Σ1/2E[Z]+μ=μCov(X,X)=E[XXT]E[X]E[X]T=Σ1/2Cov(Z)(Σ1/2)T=Σ1/2In(Σ1/2)T=ΣM(t)=E[exp(tT(ΣTZ+μ))]=E[exp(tTΣ1/2Z)]E[exp(tTμ)]=E[exp((tTΣ1/2)Z)]exp(tTμ)=exp(12tTΣt)exp(tTμ)=exp(tTμ+12tTΣt)

which must be familiar from the univariate case where mgf is exp(μt+1/2σ2t2).

The distribution of X can be obtained using the Jacobian J=dZdX=|Σ1/2|Z=Σ1/2(Xμ) where the negative sign denotes the inverse.

fX(x)=(12π)n2exp(12(()Xμ)T(Σ1/2)TΣ1/2(Xμ))|Σ1/2|=(12π)n21|Σ1/2|exp(12(Xμ)TΣ(Xμ))