Normal Distribution | Gaussian | Normal random variables (2024)

previous

next

Video Available


4.2.3 Normal (Gaussian) Distribution

The normal distribution is by far the most important probability distribution. One of the main reasonsfor that is the Central Limit Theorem (CLT) that we will discuss later in the book. To give youan idea, the CLT states that if you add a large number of random variables, the distribution of the sumwill be approximately normal under certain conditions. The importance of this result comes from the factthat many random variables in real life can be expressed as the sum of a large number of random variablesand, by the CLT, we can argue that distribution of the sum should be normal. The CLT is one of the mostimportant results in probability and we will discuss it later on. Here, we will introduce normal randomvariables.

We first define the standard normal random variable. We will then see that we can obtain other normalrandom variables by scaling and shifting a standard normal random variable.

A continuous random variable $Z$ is said to be a standard normal (standard Gaussian) random variable,shown as $Z \sim N(0,1)$, if its PDF is given by$$f_Z(z) = \frac{1}{\sqrt{2 \pi}} \exp\left\{-\frac{z^2}{2}\right\}, \hspace{20pt} \textrm{for all } z \in \mathbb{R}.$$


The $\frac{1}{\sqrt{2 \pi}}$ is there to make sure that the area under the PDF is equal to one. We willverify that this holds in the solved problems section. Figure 4.6 shows the PDF of the standard normalrandom variable.

Normal Distribution | Gaussian | Normal random variables (2)

Let us find the mean and variance of the standard normal distribution. To do that, we will use a simpleuseful fact. Consider a function $g(u):\mathbb{R}\rightarrow\mathbb{R}$. If $g(u)$ is an odd function,i.e., $g(-u)=-g(u)$, and $|\int_{0}^{\infty} g(u) du| < \infty$, then$$\int_{-\infty}^{\infty} g(u) du=0.$$For our purpose, let$$g(u)= u^{2k+1}\exp\left\{-\frac{u^2}{2}\right\},$$where $k=0,1,2,...$. Then $g(u)$ is an odd function. Also $|\int_{0}^{\infty} g(u) du| < \infty$. One wayto see this is to note that $g(u)$ decays faster than the function $\exp\left\{-u\right\}$ and since$|\int_{0}^{\infty} \exp\left\{-u\right\} du| < \infty$, we conclude that $|\int_{0}^{\infty} g(u) du| < \infty$. Now,let $Z$ be a standard normal random variable. Then, we have$$EZ^{2k+1} = \frac{1}{\sqrt{2 \pi}} \int_{-\infty}^{\infty} u^{2k+1}\exp\left\{-\frac{u^2}{2}\right\} du=0,$$for all $k \in \{0,1,2,..,\}$. Thus, we have shown that for a standard normal random variable $Z$, we have$$EZ=EZ^3=EZ^5=....=0.$$In particular, the standard normal distribution has zero mean. This is not surprising as we can seefrom Figure 4.6 that the PDF is symmetric around the origin, so we expect that $EZ=0$. Next, let's find $EZ^2$.

$EZ^2$$= \frac{1}{\sqrt{2 \pi}} \int_{-\infty}^{\infty} u^2\exp\left\{-\frac{u^2}{2}\right\} du$
$= \frac{1}{\sqrt{2 \pi}}\bigg[ -u\exp\left\{-\frac{u^2}{2}\right\}\bigg]_{-\infty}^{\infty} +$
$+\frac{1}{\sqrt{2 \pi}} \int_{-\infty}^{\infty} \exp\left\{-\frac{u^2}{2}\right\} du \hspace{20pt} (\textrm{integration by parts})\\$
$= \int_{-\infty}^{\infty} \frac{1}{\sqrt{2 \pi}} \exp\left\{-\frac{u^2}{2}\right\} du$
$=1.$

The last equality holds because we are integrating the standard normal PDF from $-\infty$ to $\infty$. Thus,we conclude that for a standard normal random variable $Z$, we have$$\textrm{Var}(Z)=1.$$So far we have shown the following:

If $Z \sim N(0,1)$, then $EZ=0$ and Var$(Z)=1$.


CDF of the standard normal

To find the CDF of the standard normal distribution, we need to integrate the PDF function.In particular, we have$$F_Z(z)=\frac{1}{\sqrt{2 \pi}} \int_{-\infty}^{z}\exp\left\{-\frac{u^2}{2}\right\} du.$$This integral does not have a closed form solution. Nevertheless, because of the importance ofthe normal distribution, the values of $F_Z(z)$ have been tabulated and many calculators andsoftware packages have this function. We usually denote the standard normal CDF by $\Phi$.

The CDF of the standard normal distribution is denoted by the $\Phi$ function:$$\Phi(x)=P(Z \leq x)= \frac{1}{\sqrt{2 \pi}} \int_{-\infty}^{x}\exp\left\{-\frac{u^2}{2}\right\} du.$$


As we will see in a moment, the CDF of any normal random variable can be written in terms of the$\Phi$ function, so the $\Phi$ function is widely used in probability. Figure 4.7 shows the $\Phi$ function.

Normal Distribution | Gaussian | Normal random variables (3)

Here are some properties of the $\Phi$ function that can be shown from its definition.

  1. $\lim \limits_{x\rightarrow \infty} \Phi(x)=1, \hspace{5pt} \lim \limits_{x\rightarrow -\infty} \Phi(x)=0$;
  2. $\Phi(0)=\frac{1}{2}$;
  3. $\Phi(-x)=1-\Phi(x)$, for all $x \in \mathbb{R}$.

Also, since the $\Phi$ function does not have a closed form, it is sometimes useful to use upper or lowerbounds. In particular we can state the following bounds (see Problem 7 in the Solved Problems section).For all $x \geq 0$,$$\hspace{50pt} \frac{1}{\sqrt{2\pi}} \frac{x}{x^2+1} \exp\left\{-\frac{x^2}{2}\right\} \leq 1-\Phi(x) \leq\frac{1}{\sqrt{2\pi}} \frac{1}{x} \exp\left\{-\frac{x^2}{2}\right\} \hspace{50pt} (4.7)$$

As we mentioned earlier, because of the importance of the normal distribution, the valuesof the $\Phi$ function have been tabulated and many calculators and software packages have thisfunction. For example, you can use the normcdf command in MATLAB to compute $\Phi(x)$ for agiven number $x$. More specifically, $normcdf(x)$ returns $\Phi(x)$. Also, the function $norminv$ returns$\Phi^{−1}(x)$. That is, if you run $x=norminv(y)$, then $x$ will be the real number for which $\Phi(x) = y$.


Normal random variables

Now that we have seen the standard normal random variable, we can obtain any normal random variableby shifting and scaling a standard normal random variable. In particular, define$$X=\sigma Z+\mu, \hspace{20pt} \textrm{where }\sigma > 0.$$Then$$EX=\sigma EZ+\mu=\mu,$$$$\textrm{Var}(X)=\sigma^2 \textrm{Var}(Z)=\sigma^2.$$We say that $X$ is a normal random variable with mean $\mu$ and variance $\sigma^2$. We write$X \sim N(\mu, \sigma^2)$.

If $Z$ is a standard normal random variable and $X=\sigma Z+\mu$, then $X$ is a normal random variablewith mean $\mu$ and variance $\sigma^2$, i.e,$$X \sim N(\mu, \sigma^2).$$


Conversely, if $X \sim N(\mu, \sigma^2)$, the random variable defined by $Z=\frac{X-\mu}{\sigma}$ is astandard normal random variable, i.e., $Z \sim N(0,1)$. To find the CDF of $X \sim N(\mu, \sigma^2)$, we can write

$F_X(x)$$=P(X \leq x)$
$=P( \sigma Z+\mu \leq x) \hspace{20pt} \big(\textrm{where }Z \sim N(0,1)\big)$
$=P\left(Z \leq \frac{x-\mu}{\sigma}\right)$
$=\Phi\left(\frac{x-\mu}{\sigma}\right).$

To find the PDF, we can take the derivative of $F_X$,

$f_X(x)$$=\frac{d}{dx} F_X(x)$
$=\frac{d}{dx} \Phi\left(\frac{x-\mu}{\sigma}\right)$
$=\frac{1}{\sigma} \Phi'\left(\frac{x-\mu}{\sigma}\right) \hspace{20pt} \textrm{(chain rule for derivative)}$
$=\frac{1}{\sigma} f_Z\left(\frac{x-\mu}{\sigma}\right)$
$=\frac{1}{\sigma\sqrt{2 \pi} } \exp\left\{-\frac{(x-\mu)^2}{2\sigma^2}\right\}.$

If $X$ is a normal random variable with mean $\mu$ and variance $\sigma^2$, i.e, $X \sim N(\mu, \sigma^2)$, then$$f_X(x)=\frac{1}{ \sigma\sqrt{2 \pi}} \exp\left\{-\frac{(x-\mu)^2}{2\sigma^2}\right\},$$$$F_X(x)=P(X \leq x)=\Phi\left(\frac{x-\mu}{\sigma}\right),$$$$P(a < X \leq b)= \Phi\left(\frac{b-\mu}{\sigma}\right)-\Phi\left(\frac{a-\mu}{\sigma}\right).$$


Figure 4.8 shows the PDF of the normal distribution for several values of $\mu$ and $\sigma$.

Normal Distribution | Gaussian | Normal random variables (4)

Example
Let $X \sim N(-5,4)$.

  1. Find $P(X < 0)$.
  2. Find $P(-7 < X < -3)$.
  3. Find $P(X > -3 | X >-5)$.
  • Solution
    • $X$ is a normal random variable with $\mu=-5$ and $\sigma=\sqrt{4}=2$, thus we have

      1. Find $P(X < 0)$:
        $P(X < 0)$$=F_X(0)$
        $=\Phi\bigg(\frac{0-(-5)}{2}\bigg)$
        $=\Phi(2.5)\approx 0.99$

      2. Find $P(-7 < X < -3)$:
        $P(-7 < X < -3)$$=F_X(-3)-F_X(-7)$
        $=\Phi\bigg(\frac{(-3)-(-5)}{2}\bigg)-\Phi\bigg(\frac{(-7)-(-5)}{2}\bigg)$
        $=\Phi(1)-\Phi(-1)$
        $=2\Phi(1)-1 \hspace{20pt} \big(\textrm{since }\Phi(-x)=1-\Phi(x)\big)$
        $\approx 0.68$

      3. Find $P(X > -3 | X > -5)$:
        $P(X > -3 | X > -5)$$=\frac{P(X > -3,X > -5)}{P(X > -5)}$
        $=\frac{P(X > -3)}{P(X > -5)}$
        $=\frac{1-\Phi\bigg(\frac{(-3)-(-5)}{2}\bigg)}{1-\Phi\bigg(\frac{(-5)-(-5)}{2}\bigg)}$
        $=\frac{1-\Phi(1)}{1-\Phi(0)}$
        $\approx \frac{0.1587}{0.5} \approx 0.32$


An important and useful property of the normal distribution is that a linear transformation of a normalrandom variable is itself a normal random variable. In particular, we have the following theorem:

Theorem
If $X \sim N(\mu_X, \sigma_X^2)$, and $Y=aX+b$, where $a,b \in \mathbb{R}$, then $Y \sim N(\mu_Y, \sigma_Y^2)$ where$$\mu_Y=a\mu_X+b, \hspace{10pt} \sigma^2_Y=a^2 \sigma_X^2.$$


Proof

We can write$$X =\sigma_X Z+ \mu_X \hspace{20pt} \textrm{where } Z \sim N(0,1).$$Thus,

$Y$$=aX+b$
$=a(\sigma_X Z+ \mu_X)+b$
$=(a \sigma_X) Z+ (a\mu_X+b).$

Therefore,$$Y \sim N(a\mu_X+b, a^2 \sigma^2_X).$$

previous

next


The print version of the book is available on Amazon.

Normal Distribution | Gaussian | Normal random variables (5)


Practical uncertainty: Useful Ideas in Decision-Making, Risk, Randomness, & AI

Normal Distribution | Gaussian | Normal random variables (6)

Normal Distribution | Gaussian | Normal random variables (2024)
Top Articles
Latest Posts
Article information

Author: Stevie Stamm

Last Updated:

Views: 6329

Rating: 5 / 5 (80 voted)

Reviews: 87% of readers found this page helpful

Author information

Name: Stevie Stamm

Birthday: 1996-06-22

Address: Apt. 419 4200 Sipes Estate, East Delmerview, WY 05617

Phone: +342332224300

Job: Future Advertising Analyst

Hobby: Leather crafting, Puzzles, Leather crafting, scrapbook, Urban exploration, Cabaret, Skateboarding

Introduction: My name is Stevie Stamm, I am a colorful, sparkling, splendid, vast, open, hilarious, tender person who loves writing and wants to share my knowledge and understanding with you.